code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: int=3 , UpperCamelCase__: Union[str, Any]=32 , UpperCamelCase__: Any=3 , UpperCamelCase__: Optional[int]=10 , UpperCamelCase__: List[str]=[10, 20, 30, 40] , UpperCamelCase__: Tuple=[1, 1, 2, 1] , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: List[str]=True , UpperCamelCase__: str="relu" , UpperCamelCase__: Any=3 , UpperCamelCase__: Optional[int]=None , ): lowerCamelCase__ : List[str] = parent lowerCamelCase__ : str = batch_size lowerCamelCase__ : str = image_size lowerCamelCase__ : List[str] = num_channels lowerCamelCase__ : List[str] = embeddings_size lowerCamelCase__ : Dict = hidden_sizes lowerCamelCase__ : Optional[Any] = depths lowerCamelCase__ : Dict = is_training lowerCamelCase__ : str = use_labels lowerCamelCase__ : Any = hidden_act lowerCamelCase__ : str = num_labels lowerCamelCase__ : List[str] = scope lowerCamelCase__ : Optional[int] = len(UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : int = None if self.use_labels: lowerCamelCase__ : Dict = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : str = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: int ): lowerCamelCase__ : Union[str, Any] = TFResNetModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: List[Any] , UpperCamelCase__: List[str] , UpperCamelCase__: int ): lowerCamelCase__ : Union[str, Any] = self.num_labels lowerCamelCase__ : Dict = TFResNetForImageClassification(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = config_and_inputs lowerCamelCase__ : Union[str, Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () a = ( {"""feature-extraction""": TFResNetModel, """image-classification""": TFResNetForImageClassification} if is_tf_available() else {} ) a = False a = False a = False a = False a = False def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Union[str, Any] = TFResNetModelTester(self ) lowerCamelCase__ : Any = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase_ ( self: int ): return @unittest.skip(reason="""ResNet does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""ResNet does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Union[str, Any] ): pass def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[Any] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[str] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Union[str, Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): def check_hidden_states_output(UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : List[str] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCamelCase__ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase__ ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: lowerCamelCase__ : Tuple = layer_type lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Union[str, Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[Any] ): for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : List[str] = TFResNetModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: int ): return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[str] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowerCamelCase__ : Union[str, Any] = self.default_image_processor lowerCamelCase__ : Optional[Any] = prepare_img() lowerCamelCase__ : Optional[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # forward pass lowerCamelCase__ : Optional[int] = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[Any] = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : Tuple = tf.constant([-11.1_069, -9.7_877, -8.3_777] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , UpperCamelCase__ , atol=1e-4 ) )
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> list[int]: if length <= 0 or not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("""Length must be a positive integer.""" ) return [n * (2 * n - 1) for n in range(UpperCamelCase )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
41
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup _A : Dict =logging.get_logger(__name__) class _lowercase ( _lowercase ): def __init__( self: Dict , **UpperCamelCase__: str ): requires_backends(self , ["""bs4"""] ) super().__init__(**UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : Any = [] lowerCamelCase__ : Tuple = [] lowerCamelCase__ : Union[str, Any] = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag lowerCamelCase__ : Optional[int] = parent.find_all(child.name , recursive=UpperCamelCase__ ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(UpperCamelCase__ ) else next(i for i, s in enumerate(UpperCamelCase__ , 1 ) if s is child ) ) lowerCamelCase__ : Dict = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : Union[str, Any] = BeautifulSoup(UpperCamelCase__ , """html.parser""" ) lowerCamelCase__ : Tuple = [] lowerCamelCase__ : Optional[Any] = [] lowerCamelCase__ : str = [] for element in html_code.descendants: if type(UpperCamelCase__ ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue lowerCamelCase__ : int = html.unescape(UpperCamelCase__ ).strip() if not text_in_this_tag: continue all_doc_strings.append(UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : Tuple = self.xpath_soup(UpperCamelCase__ ) stringaxtag_seq.append(UpperCamelCase__ ) stringaxsubs_seq.append(UpperCamelCase__ ) if len(UpperCamelCase__ ) != len(UpperCamelCase__ ): raise ValueError("""Number of doc strings and xtags does not correspond""" ) if len(UpperCamelCase__ ) != len(UpperCamelCase__ ): raise ValueError("""Number of doc strings and xsubs does not correspond""" ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def lowerCamelCase_ ( self: int , UpperCamelCase__: List[Any] , UpperCamelCase__: List[Any] ): lowerCamelCase__ : List[Any] = """""" for tagname, subs in zip(UpperCamelCase__ , UpperCamelCase__ ): xpath += F'''/{tagname}''' if subs != 0: xpath += F'''[{subs}]''' return xpath def __call__( self: List[str] , UpperCamelCase__: Tuple ): lowerCamelCase__ : List[Any] = False # Check that strings has a valid type if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : List[Any] = True elif isinstance(UpperCamelCase__ , (list, tuple) ): if len(UpperCamelCase__ ) == 0 or isinstance(html_strings[0] , UpperCamelCase__ ): lowerCamelCase__ : str = True if not valid_strings: raise ValueError( """HTML strings must of type `str`, `List[str]` (batch of examples), """ F'''but is of type {type(UpperCamelCase__ )}.''' ) lowerCamelCase__ : List[Any] = bool(isinstance(UpperCamelCase__ , (list, tuple) ) and (isinstance(html_strings[0] , UpperCamelCase__ )) ) if not is_batched: lowerCamelCase__ : int = [html_strings] # Get nodes + xpaths lowerCamelCase__ : Dict = [] lowerCamelCase__ : int = [] for html_string in html_strings: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = self.get_three_from_single(UpperCamelCase__ ) nodes.append(UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = [] for node, tag_list, sub_list in zip(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : int = self.construct_xpath(UpperCamelCase__ , UpperCamelCase__ ) xpath_strings.append(UpperCamelCase__ ) xpaths.append(UpperCamelCase__ ) # return as Dict lowerCamelCase__ : int = {"""nodes""": nodes, """xpaths""": xpaths} lowerCamelCase__ : Optional[Any] = BatchFeature(data=UpperCamelCase__ , tensor_type=UpperCamelCase__ ) return encoded_inputs
41
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
1
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
'''simple docstring''' class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : dict[str, TrieNode] = {} # Mapping from char to TrieNode lowerCamelCase__ : List[str] = False def lowerCamelCase_ ( self: str , UpperCamelCase__: list[str] ): for word in words: self.insert(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : List[Any] = self for char in word: if char not in curr.nodes: lowerCamelCase__ : Tuple = TrieNode() lowerCamelCase__ : List[Any] = curr.nodes[char] lowerCamelCase__ : Any = True def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = self for char in word: if char not in curr.nodes: return False lowerCamelCase__ : Any = curr.nodes[char] return curr.is_leaf def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): def _delete(UpperCamelCase__: TrieNode , UpperCamelCase__: str , UpperCamelCase__: int ) -> bool: if index == len(UpperCamelCase__ ): # If word does not exist if not curr.is_leaf: return False lowerCamelCase__ : str = False return len(curr.nodes ) == 0 lowerCamelCase__ : List[str] = word[index] lowerCamelCase__ : Dict = curr.nodes.get(UpperCamelCase__ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCamelCase__ : List[Any] = _delete(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCamelCase__ , 0 ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: if node.is_leaf: print(UpperCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(UpperCamelCase , word + key ) def SCREAMING_SNAKE_CASE_ () -> bool: lowerCamelCase__ : str = """banana bananas bandana band apple all beast""".split() lowerCamelCase__ : Union[str, Any] = TrieNode() root.insert_many(UpperCamelCase ) # print_words(root, "") assert all(root.find(UpperCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: print(str(UpperCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def SCREAMING_SNAKE_CASE_ () -> None: assert test_trie() def SCREAMING_SNAKE_CASE_ () -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
41
1
'''simple docstring''' import os from math import logaa def SCREAMING_SNAKE_CASE_ (UpperCamelCase = "base_exp.txt" ) -> int: lowerCamelCase__ : float = 0 lowerCamelCase__ : str = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(UpperCamelCase ) , UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : Dict = list(map(UpperCamelCase , line.split(""",""" ) ) ) if x * logaa(UpperCamelCase ) > largest: lowerCamelCase__ : Tuple = x * logaa(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = i + 1 return result if __name__ == "__main__": print(solution())
41
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' from functools import lru_cache def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> set: lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : str = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(UpperCamelCase ) if n > 1: factors.add(UpperCamelCase ) return factors @lru_cache def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: return len(unique_prime_factors(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> bool: return len(set(UpperCamelCase ) ) in (0, 1) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> list: lowerCamelCase__ : Any = 2 while True: # Increment each value of a generated range lowerCamelCase__ : List[Any] = [base + i for i in range(UpperCamelCase )] # Run elements through out unique_prime_factors function # Append our target number to the end. lowerCamelCase__ : Dict = [upf_len(UpperCamelCase ) for x in group] checker.append(UpperCamelCase ) # If all numbers in the list are equal, return the group variable. if equality(UpperCamelCase ): return group # Increment our base variable by 1 base += 1 def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 4 ) -> int: lowerCamelCase__ : List[Any] = run(UpperCamelCase ) return results[0] if len(UpperCamelCase ) else None if __name__ == "__main__": print(solution())
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: int , UpperCamelCase__: Dict=32 ): set_seed(0 ) lowerCamelCase__ : str = UNetaDModel(sample_size=UpperCamelCase__ , in_channels=3 , out_channels=3 ) lowerCamelCase__ : Dict = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Any = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowerCamelCase__ : int = DDPMScheduler( num_train_timesteps=1_000 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=UpperCamelCase__ , ) lowerCamelCase__ : int = DDIMScheduler( num_train_timesteps=1_000 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=UpperCamelCase__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowerCamelCase__ : Tuple = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(UpperCamelCase__ ) for _ in range(4 )] lowerCamelCase__ : str = [torch.randn((4, 3, 32, 32) ).to(UpperCamelCase__ ) for _ in range(4 )] lowerCamelCase__ : Tuple = [torch.randint(0 , 1_000 , (4,) ).long().to(UpperCamelCase__ ) for _ in range(4 )] # train with a DDPM scheduler lowerCamelCase__ , lowerCamelCase__ : int = self.get_model_optimizer(resolution=32 ) model.train().to(UpperCamelCase__ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase__ : Any = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , timesteps[i] ).sample lowerCamelCase__ : Union[str, Any] = torch.nn.functional.mse_loss(UpperCamelCase__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowerCamelCase__ , lowerCamelCase__ : Any = self.get_model_optimizer(resolution=32 ) model.train().to(UpperCamelCase__ ) for i in range(4 ): optimizer.zero_grad() lowerCamelCase__ : Union[str, Any] = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , timesteps[i] ).sample lowerCamelCase__ : int = torch.nn.functional.mse_loss(UpperCamelCase__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-5 ) )
41
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _A : List[Any] =logging.get_logger(__name__) _A : Dict =['''model.decoder.embed_positions.weights'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: if "emb" in name: lowerCamelCase__ : Dict = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: lowerCamelCase__ : List[str] = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: lowerCamelCase__ : List[str] = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: lowerCamelCase__ : Optional[int] = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: lowerCamelCase__ : Dict = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: lowerCamelCase__ : Optional[Any] = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: lowerCamelCase__ : Optional[Any] = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: lowerCamelCase__ : int = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple[Dict, Dict]: lowerCamelCase__ : int = list(state_dict.keys() ) lowerCamelCase__ : Tuple = {} for key in keys: lowerCamelCase__ : Any = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj lowerCamelCase__ : Union[str, Any] = val[:hidden_size, :] lowerCamelCase__ : Any = val[hidden_size : 2 * hidden_size, :] lowerCamelCase__ : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: lowerCamelCase__ : str = val else: lowerCamelCase__ : Union[str, Any] = val return state_dict, enc_dec_proj_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values lowerCamelCase__ : int = 1024 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 16 elif checkpoint == "medium": lowerCamelCase__ : Any = 1536 lowerCamelCase__ : Union[str, Any] = 48 lowerCamelCase__ : Optional[int] = 24 elif checkpoint == "large": lowerCamelCase__ : Optional[Any] = 2048 lowerCamelCase__ : Dict = 48 lowerCamelCase__ : List[Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) lowerCamelCase__ : Any = MusicgenDecoderConfig( hidden_size=UpperCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase , num_attention_heads=UpperCamelCase , ) return config @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase="cpu" ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = MusicGen.get_pretrained(UpperCamelCase , device=UpperCamelCase ) lowerCamelCase__ : List[Any] = decoder_config_from_checkpoint(UpperCamelCase ) lowerCamelCase__ : Any = fairseq_model.lm.state_dict() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = rename_state_dict( UpperCamelCase , hidden_size=decoder_config.hidden_size ) lowerCamelCase__ : str = TaEncoderModel.from_pretrained("""t5-base""" ) lowerCamelCase__ : Tuple = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) lowerCamelCase__ : Optional[int] = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection lowerCamelCase__ , lowerCamelCase__ : List[str] = decoder.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model lowerCamelCase__ : Optional[Any] = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase , audio_encoder=UpperCamelCase , decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass lowerCamelCase__ : Dict = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) lowerCamelCase__ : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor lowerCamelCase__ : str = AutoTokenizer.from_pretrained("""t5-base""" ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) lowerCamelCase__ : Optional[int] = MusicgenProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids lowerCamelCase__ : Union[str, Any] = 2048 lowerCamelCase__ : List[str] = 2048 # set other default generation config params lowerCamelCase__ : Optional[Any] = int(30 * audio_encoder.config.frame_rate ) lowerCamelCase__ : Union[str, Any] = True lowerCamelCase__ : List[Any] = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _A : List[str] =parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
41
1
'''simple docstring''' from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging _A : int =logging.get_logger(__name__) class _lowercase : a = 42 a = None @staticmethod def lowerCamelCase_ ( ): raise NotImplementedError def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: int , UpperCamelCase__: str , **UpperCamelCase__: Union[str, Any] ): raise NotImplementedError def lowerCamelCase_ ( self: Any , UpperCamelCase__: Any ): raise NotImplementedError def lowerCamelCase_ ( self: List[Any] ): if not self.is_available(): raise RuntimeError( F'''You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}.''' ) @classmethod def lowerCamelCase_ ( cls: Optional[int] ): return F'''`pip install {cls.pip_package or cls.name}`''' class _lowercase ( _lowercase ): a = """optuna""" @staticmethod def lowerCamelCase_ ( ): return is_optuna_available() def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: int , UpperCamelCase__: str , **UpperCamelCase__: Dict ): return run_hp_search_optuna(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Any ): return default_hp_space_optuna(UpperCamelCase__ ) class _lowercase ( _lowercase ): a = """ray""" a = """'ray[tune]'""" @staticmethod def lowerCamelCase_ ( ): return is_ray_available() def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: str , UpperCamelCase__: int , UpperCamelCase__: str , **UpperCamelCase__: Union[str, Any] ): return run_hp_search_ray(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: Any ): return default_hp_space_ray(UpperCamelCase__ ) class _lowercase ( _lowercase ): a = """sigopt""" @staticmethod def lowerCamelCase_ ( ): return is_sigopt_available() def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: int , UpperCamelCase__: str , **UpperCamelCase__: Dict ): return run_hp_search_sigopt(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[str] ): return default_hp_space_sigopt(UpperCamelCase__ ) class _lowercase ( _lowercase ): a = """wandb""" @staticmethod def lowerCamelCase_ ( ): return is_wandb_available() def lowerCamelCase_ ( self: int , UpperCamelCase__: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , **UpperCamelCase__: int ): return run_hp_search_wandb(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Union[str, Any] ): return default_hp_space_wandb(UpperCamelCase__ ) _A : Dict ={ HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def SCREAMING_SNAKE_CASE_ () -> str: lowerCamelCase__ : List[str] = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(UpperCamelCase ) > 0: lowerCamelCase__ : Dict = available_backends[0].name if len(UpperCamelCase ) > 1: logger.info( f'''{len(UpperCamelCase )} hyperparameter search backends available. Using {name} as the default.''' ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( f''' - To install {backend.name} run {backend.pip_install()}''' for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
41
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = array[indexa], array[indexa] def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if length > 1: lowerCamelCase__ : Optional[Any] = int(length / 2 ) for i in range(UpperCamelCase , low + middle ): comp_and_swap(UpperCamelCase , UpperCamelCase , i + middle , UpperCamelCase ) bitonic_merge(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) bitonic_merge(UpperCamelCase , low + middle , UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if length > 1: lowerCamelCase__ : Tuple = int(length / 2 ) bitonic_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , 1 ) bitonic_sort(UpperCamelCase , low + middle , UpperCamelCase , 0 ) bitonic_merge(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": _A : List[str] =input('''Enter numbers separated by a comma:\n''').strip() _A : List[str] =[int(item.strip()) for item in user_input.split(''',''')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('''\nSorted array in ascending order is: ''', end='''''') print(*unsorted, sep=''', ''') bitonic_merge(unsorted, 0, len(unsorted), 0) print('''Sorted array in descending order is: ''', end='''''') print(*unsorted, sep=''', ''')
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features _A : List[str] =logging.get_logger(__name__) _A : Optional[int] =list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) _A : str =tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _lowercase : a = field( default=_lowercase , metadata={"""help""": """Model type selected in the list: """ + """, """.join(_lowercase )} ) a = field( default=_lowercase , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) a = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) a = field( default=128 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) a = field( default=64 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) a = field( default=30 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) a = field( default=_lowercase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) a = field( default=_lowercase , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) a = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) a = field( default=20 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) a = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) a = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class _lowercase ( _lowercase ): a = """train""" a = """dev""" class _lowercase ( _lowercase ): a = 42 a = 42 a = 42 a = 42 def __init__( self: Optional[Any] , UpperCamelCase__: SquadDataTrainingArguments , UpperCamelCase__: PreTrainedTokenizer , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: Union[str, Split] = Split.train , UpperCamelCase__: Optional[bool] = False , UpperCamelCase__: Optional[str] = None , UpperCamelCase__: Optional[str] = "pt" , ): lowerCamelCase__ : Optional[Any] = args lowerCamelCase__ : Optional[int] = is_language_sensitive lowerCamelCase__ : Optional[int] = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(UpperCamelCase__ , UpperCamelCase__ ): try: lowerCamelCase__ : Optional[int] = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) lowerCamelCase__ : str = mode # Load data features from cache or dataset file lowerCamelCase__ : str = """v2""" if args.version_2_with_negative else """v1""" lowerCamelCase__ : int = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : Optional[Any] = cached_features_file + """.lock""" with FileLock(UpperCamelCase__ ): if os.path.exists(UpperCamelCase__ ) and not args.overwrite_cache: lowerCamelCase__ : Tuple = time.time() lowerCamelCase__ : Union[str, Any] = torch.load(UpperCamelCase__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowerCamelCase__ : int = self.old_features["""features"""] lowerCamelCase__ : Optional[Any] = self.old_features.get("""dataset""" , UpperCamelCase__ ) lowerCamelCase__ : Any = self.old_features.get("""examples""" , UpperCamelCase__ ) logger.info( F'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( F'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' """ future run""" ) else: if mode == Split.dev: lowerCamelCase__ : List[Any] = self.processor.get_dev_examples(args.data_dir ) else: lowerCamelCase__ : Optional[int] = self.processor.get_train_examples(args.data_dir ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=UpperCamelCase__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=UpperCamelCase__ , ) lowerCamelCase__ : int = time.time() torch.save( {"""features""": self.features, """dataset""": self.dataset, """examples""": self.examples} , UpperCamelCase__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self: Optional[int] ): return len(self.features ) def __getitem__( self: int , UpperCamelCase__: List[Any] ): # Convert to Tensors and build dataset lowerCamelCase__ : Any = self.features[i] lowerCamelCase__ : Union[str, Any] = torch.tensor(feature.input_ids , dtype=torch.long ) lowerCamelCase__ : List[str] = torch.tensor(feature.attention_mask , dtype=torch.long ) lowerCamelCase__ : Any = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowerCamelCase__ : int = torch.tensor(feature.cls_index , dtype=torch.long ) lowerCamelCase__ : Optional[Any] = torch.tensor(feature.p_mask , dtype=torch.float ) lowerCamelCase__ : Any = torch.tensor(feature.is_impossible , dtype=torch.float ) lowerCamelCase__ : int = { """input_ids""": input_ids, """attention_mask""": attention_mask, """token_type_ids""": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"""cls_index""": cls_index, """p_mask""": p_mask} ) if self.args.version_2_with_negative: inputs.update({"""is_impossible""": is_impossible} ) if self.is_language_sensitive: inputs.update({"""langs""": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowerCamelCase__ : Any = torch.tensor(feature.start_position , dtype=torch.long ) lowerCamelCase__ : Optional[Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"""start_positions""": start_positions, """end_positions""": end_positions} ) return inputs
41
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=7 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[str]=18 , UpperCamelCase__: Union[str, Any]=30 , UpperCamelCase__: List[str]=400 , UpperCamelCase__: Any=True , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[Any]=False , UpperCamelCase__: Tuple=True , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , ): lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Optional[int] = min_resolution lowerCamelCase__ : Optional[Any] = max_resolution lowerCamelCase__ : Union[str, Any] = do_resize lowerCamelCase__ : Optional[Any] = size if size is not None else {"""height""": 18, """width""": 20} lowerCamelCase__ : Dict = do_thumbnail lowerCamelCase__ : Optional[int] = do_align_axis lowerCamelCase__ : Any = do_pad lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Union[str, Any] = image_mean lowerCamelCase__ : Union[str, Any] = image_std def lowerCamelCase_ ( self: str ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DonutImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = DonutImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_thumbnail""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_align_long_axis""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 20} ) lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order lowerCamelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"""height""": 84, """width""": 42} ) def lowerCamelCase_ ( self: List[str] ): pass @is_flaky() def lowerCamelCase_ ( self: Union[str, Any] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Optional[int] ): # Initialize image_processing lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Dict ): # Initialize image_processing lowerCamelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
41
1
'''simple docstring''' import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() _A : Optional[Any] =logging.get_logger(__name__) _A : int ={name: getattr(transformers, name + '''Fast''') for name in SLOW_TO_FAST_CONVERTERS} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[int]: if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(f'''Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.''' ) if tokenizer_name is None: lowerCamelCase__ : Optional[Any] = TOKENIZER_CLASSES else: lowerCamelCase__ : Dict = {tokenizer_name: getattr(UpperCamelCase , tokenizer_name + """Fast""" )} logger.info(f'''Loading tokenizer classes: {tokenizer_names}''' ) for tokenizer_name in tokenizer_names: lowerCamelCase__ : Tuple = TOKENIZER_CLASSES[tokenizer_name] lowerCamelCase__ : Optional[int] = True if checkpoint_name is None: lowerCamelCase__ : Any = list(tokenizer_class.max_model_input_sizes.keys() ) else: lowerCamelCase__ : Any = [checkpoint_name] logger.info(f'''For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}''' ) for checkpoint in checkpoint_names: logger.info(f'''Loading {tokenizer_class.__class__.__name__} {checkpoint}''' ) # Load tokenizer lowerCamelCase__ : List[str] = tokenizer_class.from_pretrained(UpperCamelCase , force_download=UpperCamelCase ) # Save fast tokenizer logger.info(f'''Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}''' ) # For organization names we create sub-directories if "/" in checkpoint: lowerCamelCase__ , lowerCamelCase__ : List[str] = checkpoint.split("""/""" ) lowerCamelCase__ : Union[str, Any] = os.path.join(UpperCamelCase , UpperCamelCase ) elif add_prefix: lowerCamelCase__ : Optional[int] = checkpoint lowerCamelCase__ : Optional[Any] = dump_path else: lowerCamelCase__ : str = None lowerCamelCase__ : Union[str, Any] = dump_path logger.info(f'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: lowerCamelCase__ : Tuple = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] lowerCamelCase__ : Optional[Any] = file_path.split(UpperCamelCase )[-1][0] if next_char == "/": lowerCamelCase__ : Optional[Any] = os.path.join(UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : List[Any] = None logger.info(f'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) lowerCamelCase__ : Any = tokenizer.save_pretrained( UpperCamelCase , legacy_format=UpperCamelCase , filename_prefix=UpperCamelCase ) logger.info(f'''=> File names {file_names}''' ) for file_name in file_names: if not file_name.endswith("""tokenizer.json""" ): os.remove(UpperCamelCase ) logger.info(f'''=> removing {file_name}''' ) if __name__ == "__main__": _A : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output generated fast tokenizer files.''' ) parser.add_argument( '''--tokenizer_name''', default=None, type=str, help=( F'Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will ' '''download and convert all the checkpoints from AWS.''' ), ) parser.add_argument( '''--checkpoint_name''', default=None, type=str, help='''Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.''', ) parser.add_argument( '''--force_download''', action='''store_true''', help='''Re-download checkpoints.''', ) _A : Optional[int] =parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger _A : Optional[int] ='''<<<<<<< This should probably be modified because it mentions: ''' _A : List[str] ='''======= >>>>>>> ''' _A : Optional[Any] =[ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] _A : List[Any] =[ # (pattern, replacement) # Order is important here for some replacements (r'''tfds\.core''', r'''datasets'''), (r'''tf\.io\.gfile\.GFile''', r'''open'''), (r'''tf\.([\w\d]+)''', r'''datasets.Value(\'\1\')'''), (r'''tfds\.features\.Text\(\)''', r'''datasets.Value(\'string\')'''), (r'''tfds\.features\.Text\(''', r'''datasets.Value(\'string\'),'''), (r'''features\s*=\s*tfds.features.FeaturesDict\(''', r'''features=datasets.Features('''), (r'''tfds\.features\.FeaturesDict\(''', r'''dict('''), (r'''The TensorFlow Datasets Authors''', r'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (r'''tfds\.''', r'''datasets.'''), (r'''dl_manager\.manual_dir''', r'''self.config.data_dir'''), (r'''self\.builder_config''', r'''self.config'''), ] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: return ConvertCommand(args.tfds_path , args.datasets_directory ) class _lowercase ( _lowercase ): @staticmethod def lowerCamelCase_ ( UpperCamelCase__: ArgumentParser ): lowerCamelCase__ : Any = parser.add_parser( """convert""" , help="""Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.""" , ) train_parser.add_argument( """--tfds_path""" , type=UpperCamelCase__ , required=UpperCamelCase__ , help="""Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.""" , ) train_parser.add_argument( """--datasets_directory""" , type=UpperCamelCase__ , required=UpperCamelCase__ , help="""Path to the HuggingFace Datasets folder.""" ) train_parser.set_defaults(func=UpperCamelCase__ ) def __init__( self: Any , UpperCamelCase__: str , UpperCamelCase__: str , *UpperCamelCase__: Any ): lowerCamelCase__ : List[str] = get_logger("""datasets-cli/converting""" ) lowerCamelCase__ : Union[str, Any] = tfds_path lowerCamelCase__ : Union[str, Any] = datasets_directory def lowerCamelCase_ ( self: Optional[int] ): if os.path.isdir(self._tfds_path ): lowerCamelCase__ : int = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowerCamelCase__ : Tuple = os.path.dirname(self._tfds_path ) else: raise ValueError("""--tfds_path is neither a directory nor a file. Please check path.""" ) lowerCamelCase__ : Optional[Any] = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) lowerCamelCase__ : Optional[Any] = [] lowerCamelCase__ : str = [] lowerCamelCase__ : str = {} if os.path.isdir(self._tfds_path ): lowerCamelCase__ : int = os.listdir(UpperCamelCase__ ) else: lowerCamelCase__ : Dict = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) lowerCamelCase__ : Union[str, Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Tuple = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) if not os.path.isfile(UpperCamelCase__ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("""Skipping file""" ) continue with open(UpperCamelCase__ , encoding="""utf-8""" ) as f: lowerCamelCase__ : Dict = f.readlines() lowerCamelCase__ : Union[str, Any] = [] lowerCamelCase__ : List[str] = False lowerCamelCase__ : List[Any] = False lowerCamelCase__ : Optional[Any] = [] for line in lines: lowerCamelCase__ : Union[str, Any] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowerCamelCase__ : str = """import datasets\n""" elif "import tensorflow" in out_line: # order is important here lowerCamelCase__ : Optional[int] = """""" continue elif "from absl import logging" in out_line: lowerCamelCase__ : Union[str, Any] = """from datasets import logging\n""" elif "getLogger" in out_line: lowerCamelCase__ : List[Any] = out_line.replace("""getLogger""" , """get_logger""" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowerCamelCase__ : str = True lowerCamelCase__ : Optional[int] = list(filter(lambda UpperCamelCase__ : e in out_line , UpperCamelCase__ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(UpperCamelCase__ ) + """\n""" ) out_lines.append(UpperCamelCase__ ) out_lines.append(UpperCamelCase__ ) continue else: for pattern, replacement in TO_CONVERT: lowerCamelCase__ : Dict = re.sub(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowerCamelCase__ : str = re.match(R"""from\stensorflow_datasets.*import\s([^\.\r\n]+)""" , UpperCamelCase__ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(""",""" ) ) lowerCamelCase__ : Tuple = """from . import """ + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowerCamelCase__ : List[Any] = True out_lines.append(UpperCamelCase__ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowerCamelCase__ : Any = f_name.replace(""".py""" , """""" ) lowerCamelCase__ : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(UpperCamelCase__ ) if needs_manual_update: with_manual_update.append(UpperCamelCase__ ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.writelines(UpperCamelCase__ ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: lowerCamelCase__ : int = os.path.basename(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = imports_to_builder_map[f_name.replace(""".py""" , """""" )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(UpperCamelCase__ , UpperCamelCase__ ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1
'''simple docstring''' import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging _A : Any =logging.get_logger(__name__) class _lowercase ( _lowercase ): a = CLIPConfig a = ["""CLIPEncoderLayer"""] def __init__( self: Union[str, Any] , UpperCamelCase__: CLIPConfig ): super().__init__(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = CLIPVisionModelWithProjection(config.vision_config ) lowerCamelCase__ : str = nn.Linear(config.vision_config.projection_dim , 1 ) lowerCamelCase__ : Any = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def lowerCamelCase_ ( self: Dict , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any]=0.5 , UpperCamelCase__: Tuple=0.5 ): lowerCamelCase__ : List[str] = self.vision_model(UpperCamelCase__ )[0] lowerCamelCase__ : Optional[Any] = self.p_head(UpperCamelCase__ ) lowerCamelCase__ : Any = nsfw_detected.flatten() lowerCamelCase__ : List[str] = nsfw_detected > p_threshold lowerCamelCase__ : int = nsfw_detected.tolist() if any(UpperCamelCase__ ): logger.warning( """Potential NSFW content was detected in one or more images. A black image will be returned instead.""" """ Try again with a different prompt and/or seed.""" ) for idx, nsfw_detected_ in enumerate(UpperCamelCase__ ): if nsfw_detected_: lowerCamelCase__ : Optional[Any] = np.zeros(images[idx].shape ) lowerCamelCase__ : List[Any] = self.w_head(UpperCamelCase__ ) lowerCamelCase__ : Any = watermark_detected.flatten() lowerCamelCase__ : Optional[int] = watermark_detected > w_threshold lowerCamelCase__ : int = watermark_detected.tolist() if any(UpperCamelCase__ ): logger.warning( """Potential watermarked content was detected in one or more images. A black image will be returned instead.""" """ Try again with a different prompt and/or seed.""" ) for idx, watermark_detected_ in enumerate(UpperCamelCase__ ): if watermark_detected_: lowerCamelCase__ : List[str] = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
41
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowerCamelCase__ : List[Any] = torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = pipe( image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowerCamelCase__ : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase__ : Tuple = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
41
1
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class _lowercase : a = 42 # setable values a = 42 a = 42 a = None @classmethod def lowerCamelCase_ ( cls: Optional[int] , UpperCamelCase__: CommonSchedulerState , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: jnp.ndarray ): return cls(common=UpperCamelCase__ , init_noise_sigma=UpperCamelCase__ , timesteps=UpperCamelCase__ ) @dataclass class _lowercase ( _lowercase ): a = 42 class _lowercase ( _lowercase , _lowercase ): a = [e.name for e in FlaxKarrasDiffusionSchedulers] a = 42 @property def lowerCamelCase_ ( self: str ): return True @register_to_config def __init__( self: List[str] , UpperCamelCase__: int = 1_000 , UpperCamelCase__: float = 0.0_001 , UpperCamelCase__: float = 0.02 , UpperCamelCase__: str = "linear" , UpperCamelCase__: Optional[jnp.ndarray] = None , UpperCamelCase__: str = "fixed_small" , UpperCamelCase__: bool = True , UpperCamelCase__: str = "epsilon" , UpperCamelCase__: jnp.dtype = jnp.floataa , ): lowerCamelCase__ : Any = dtype def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Optional[CommonSchedulerState] = None ): if common is None: lowerCamelCase__ : Any = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution lowerCamelCase__ : Optional[int] = jnp.array(1.0 , dtype=self.dtype ) lowerCamelCase__ : int = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=UpperCamelCase__ , init_noise_sigma=UpperCamelCase__ , timesteps=UpperCamelCase__ , ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: Optional[int] = None ): return sample def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: int , UpperCamelCase__: Tuple = () ): lowerCamelCase__ : Union[str, Any] = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 lowerCamelCase__ : Optional[Any] = (jnp.arange(0 , UpperCamelCase__ ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=UpperCamelCase__ , timesteps=UpperCamelCase__ , ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: Dict , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: Any=None ): lowerCamelCase__ : Optional[int] = state.common.alphas_cumprod[t] lowerCamelCase__ : Union[str, Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample lowerCamelCase__ : List[Any] = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: lowerCamelCase__ : Optional[int] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": lowerCamelCase__ : Tuple = jnp.clip(UpperCamelCase__ , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": lowerCamelCase__ : Union[str, Any] = jnp.log(jnp.clip(UpperCamelCase__ , a_min=1e-20 ) ) elif variance_type == "fixed_large": lowerCamelCase__ : Union[str, Any] = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log lowerCamelCase__ : Union[str, Any] = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": lowerCamelCase__ : Optional[int] = variance lowerCamelCase__ : List[Any] = state.common.betas[t] lowerCamelCase__ : List[Any] = (predicted_variance + 1) / 2 lowerCamelCase__ : List[Any] = frac * max_log + (1 - frac) * min_log return variance def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: int , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: Optional[jax.random.KeyArray] = None , UpperCamelCase__: bool = True , ): lowerCamelCase__ : int = timestep if key is None: lowerCamelCase__ : Optional[int] = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: lowerCamelCase__ , lowerCamelCase__ : str = jnp.split(UpperCamelCase__ , sample.shape[1] , axis=1 ) else: lowerCamelCase__ : Union[str, Any] = None # 1. compute alphas, betas lowerCamelCase__ : str = state.common.alphas_cumprod[t] lowerCamelCase__ : str = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) lowerCamelCase__ : str = 1 - alpha_prod_t lowerCamelCase__ : Union[str, Any] = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": lowerCamelCase__ : str = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": lowerCamelCase__ : Union[str, Any] = model_output elif self.config.prediction_type == "v_prediction": lowerCamelCase__ : List[str] = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' """ for the FlaxDDPMScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: lowerCamelCase__ : Dict = jnp.clip(UpperCamelCase__ , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowerCamelCase__ : Optional[int] = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t lowerCamelCase__ : Optional[int] = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowerCamelCase__ : Dict = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): lowerCamelCase__ : List[Any] = jax.random.split(UpperCamelCase__ , num=1 ) lowerCamelCase__ : str = jax.random.normal(UpperCamelCase__ , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(UpperCamelCase__ , UpperCamelCase__ , predicted_variance=UpperCamelCase__ ) ** 0.5) * noise lowerCamelCase__ : Optional[Any] = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) lowerCamelCase__ : Any = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=UpperCamelCase__ , state=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: jnp.ndarray , ): return add_noise_common(state.common , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: DDPMSchedulerState , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: jnp.ndarray , UpperCamelCase__: jnp.ndarray , ): return get_velocity_common(state.common , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def __len__( self: Dict ): return self.config.num_train_timesteps
41
'''simple docstring''' # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _A : Dict ='''tiny-wmt19-en-ru''' # Build # borrowed from a test _A : List[str] =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _A : str =dict(zip(vocab, range(len(vocab)))) _A : List[str] =['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: _A : Union[str, Any] =Path(tmpdirname) _A : str =build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] _A : int =build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] _A : List[Any] =build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) _A : int =FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _A : List[str] =FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _A : Union[str, Any] =FSMTForConditionalGeneration(config) print(F'num of params {tiny_model.num_parameters()}') # Test _A : List[str] =tokenizer(['''Making tiny model'''], return_tensors='''pt''') _A : Tuple =tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'Generated {mname_tiny}') # Upload # transformers-cli upload tiny-wmt19-en-ru
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _A : Any ={ '''configuration_data2vec_audio''': ['''DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecAudioConfig'''], '''configuration_data2vec_text''': [ '''DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecTextConfig''', '''Data2VecTextOnnxConfig''', ], '''configuration_data2vec_vision''': [ '''DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Data2VecVisionConfig''', '''Data2VecVisionOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecAudioForAudioFrameClassification''', '''Data2VecAudioForCTC''', '''Data2VecAudioForSequenceClassification''', '''Data2VecAudioForXVector''', '''Data2VecAudioModel''', '''Data2VecAudioPreTrainedModel''', ] _A : Optional[Any] =[ '''DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecTextForCausalLM''', '''Data2VecTextForMaskedLM''', '''Data2VecTextForMultipleChoice''', '''Data2VecTextForQuestionAnswering''', '''Data2VecTextForSequenceClassification''', '''Data2VecTextForTokenClassification''', '''Data2VecTextModel''', '''Data2VecTextPreTrainedModel''', ] _A : Optional[int] =[ '''DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Data2VecVisionForImageClassification''', '''Data2VecVisionForMaskedImageModeling''', '''Data2VecVisionForSemanticSegmentation''', '''Data2VecVisionModel''', '''Data2VecVisionPreTrainedModel''', ] if is_tf_available(): _A : List[Any] =[ '''TFData2VecVisionForImageClassification''', '''TFData2VecVisionForSemanticSegmentation''', '''TFData2VecVisionModel''', '''TFData2VecVisionPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : List[Any] =logging.get_logger(__name__) _A : Union[str, Any] ={ '''naver-clova-ix/donut-base''': '''https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json''', # See all Donut models at https://huggingface.co/models?filter=donut-swin } class _lowercase ( _lowercase ): a = """donut-swin""" a = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self: Dict , UpperCamelCase__: Tuple=224 , UpperCamelCase__: List[Any]=4 , UpperCamelCase__: Dict=3 , UpperCamelCase__: Optional[int]=96 , UpperCamelCase__: Optional[Any]=[2, 2, 6, 2] , UpperCamelCase__: List[str]=[3, 6, 12, 24] , UpperCamelCase__: Dict=7 , UpperCamelCase__: Optional[Any]=4.0 , UpperCamelCase__: List[Any]=True , UpperCamelCase__: Dict=0.0 , UpperCamelCase__: Union[str, Any]=0.0 , UpperCamelCase__: str=0.1 , UpperCamelCase__: Dict="gelu" , UpperCamelCase__: Union[str, Any]=False , UpperCamelCase__: Tuple=0.02 , UpperCamelCase__: int=1e-5 , **UpperCamelCase__: Tuple , ): super().__init__(**UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : List[Any] = patch_size lowerCamelCase__ : str = num_channels lowerCamelCase__ : Any = embed_dim lowerCamelCase__ : Dict = depths lowerCamelCase__ : List[Any] = len(UpperCamelCase__ ) lowerCamelCase__ : str = num_heads lowerCamelCase__ : Optional[int] = window_size lowerCamelCase__ : str = mlp_ratio lowerCamelCase__ : List[str] = qkv_bias lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : str = attention_probs_dropout_prob lowerCamelCase__ : Union[str, Any] = drop_path_rate lowerCamelCase__ : Dict = hidden_act lowerCamelCase__ : Union[str, Any] = use_absolute_embeddings lowerCamelCase__ : Tuple = layer_norm_eps lowerCamelCase__ : Dict = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCamelCase__ : Union[str, Any] = int(embed_dim * 2 ** (len(UpperCamelCase__ ) - 1) )
41
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: Optional[Any]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: str=0.6 , UpperCamelCase__: str=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Optional[Any] = batch_size lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Optional[Any] = is_training lowerCamelCase__ : int = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : List[str] = mask_ratio lowerCamelCase__ : Optional[int] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Any = (image_size // patch_size) ** 2 lowerCamelCase__ : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches lowerCamelCase__ : List[str] = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Union[str, Any] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Optional[int] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = TFViTMAEModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : int = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : str = outputs_dict[0].numpy() lowerCamelCase__ : Optional[int] = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def lowerCamelCase_ ( self: Dict ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__: int ): lowerCamelCase__ : Optional[int] = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): lowerCamelCase__ : List[str] = v.numpy() else: lowerCamelCase__ : Union[str, Any] = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = prepare_numpy_arrays(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : Any = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: str ): # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase__ : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : Tuple = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[Any] = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } lowerCamelCase__ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase__ : List[str] = main_layer_class(UpperCamelCase__ ) lowerCamelCase__ : int = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase__ : List[str] = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) lowerCamelCase__ : Union[str, Any] = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : int = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) lowerCamelCase__ : int = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : Any = outputs.last_hidden_state.numpy() lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : int = outputs.logits.numpy() lowerCamelCase__ : Dict = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) lowerCamelCase__ : Dict = model_class.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : str = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase__ : Optional[Any] = 0 else: lowerCamelCase__ : Union[str, Any] = after_outputs["""logits"""].numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) def lowerCamelCase_ ( self: Any ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Optional[int] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) lowerCamelCase__ : int = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase__ : int = model_class.from_config(model.config ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @slow def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[Any] ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: List[str] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Tuple = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase__ : str = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Any = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1e-4 )
41
1
'''simple docstring''' import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast _A : Optional[Any] =datasets.utils.logging.get_logger(__name__) @dataclass class _lowercase ( datasets.BuilderConfig ): a = 1_0000 a = None a = None class _lowercase ( datasets.ArrowBasedBuilder ): a = ParquetConfig def lowerCamelCase_ ( self: Optional[Any] ): return datasets.DatasetInfo(features=self.config.features ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Optional[int] ): if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) lowerCamelCase__ : Union[str, Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCamelCase__ , (str, list, tuple) ): lowerCamelCase__ : Any = data_files if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : Any = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive lowerCamelCase__ : Optional[int] = [dl_manager.iter_files(UpperCamelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] lowerCamelCase__ : Dict = [] for split_name, files in data_files.items(): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : Any = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive lowerCamelCase__ : List[Any] = [dl_manager.iter_files(UpperCamelCase__ ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(UpperCamelCase__ ): with open(UpperCamelCase__ , """rb""" ) as f: lowerCamelCase__ : Optional[Any] = datasets.Features.from_arrow_schema(pq.read_schema(UpperCamelCase__ ) ) break splits.append(datasets.SplitGenerator(name=UpperCamelCase__ , gen_kwargs={"""files""": files} ) ) return splits def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: pa.Table ): if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example lowerCamelCase__ : Optional[int] = table_cast(UpperCamelCase__ , self.info.features.arrow_schema ) return pa_table def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int ): lowerCamelCase__ : str = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F'''Tried to load parquet data with columns \'{self.config.columns}\' with mismatching features \'{self.info.features}\'''' ) for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCamelCase__ ) ): with open(UpperCamelCase__ , """rb""" ) as f: lowerCamelCase__ : List[Any] = pq.ParquetFile(UpperCamelCase__ ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): lowerCamelCase__ : Optional[Any] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F'''{file_idx}_{batch_idx}''', self._cast_table(UpperCamelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(UpperCamelCase__ )}: {e}''' ) raise
41
'''simple docstring''' class _lowercase : def __init__( self: Tuple , UpperCamelCase__: list[int] ): lowerCamelCase__ : Union[str, Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [0] * len_array if len_array > 0: lowerCamelCase__ : Union[str, Any] = array[0] for i in range(1 , UpperCamelCase__ ): lowerCamelCase__ : Tuple = self.prefix_sum[i - 1] + array[i] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: int ): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : Dict = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(UpperCamelCase__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> tuple[str, float]: if (stress, tangential_force, area).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif stress < 0: raise ValueError("""Stress cannot be negative""" ) elif tangential_force < 0: raise ValueError("""Tangential Force cannot be negative""" ) elif area < 0: raise ValueError("""Area cannot be negative""" ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A : Dict ={ '''configuration_swinv2''': ['''SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Swinv2Config'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Swinv2ForImageClassification''', '''Swinv2ForMaskedImageModeling''', '''Swinv2Model''', '''Swinv2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class _lowercase ( _lowercase , _lowercase ): a = 1 @register_to_config def __init__( self: List[Any] , UpperCamelCase__: Any=2_000 , UpperCamelCase__: Any=0.1 , UpperCamelCase__: Union[str, Any]=20 , UpperCamelCase__: int=1e-3 ): lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Dict = None lowerCamelCase__ : Dict = None def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Union[str, torch.device] = None ): lowerCamelCase__ : List[Any] = torch.linspace(1 , self.config.sampling_eps , UpperCamelCase__ , device=UpperCamelCase__ ) def lowerCamelCase_ ( self: int , UpperCamelCase__: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any]=None ): if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler""" ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCamelCase__ : Dict = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCamelCase__ : Optional[int] = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCamelCase__ : Dict = std.flatten() while len(std.shape ) < len(score.shape ): lowerCamelCase__ : Any = std.unsqueeze(-1 ) lowerCamelCase__ : Dict = -score / std # compute lowerCamelCase__ : Any = -1.0 / len(self.timesteps ) lowerCamelCase__ : Optional[int] = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCamelCase__ : List[str] = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCamelCase__ : Tuple = beta_t.unsqueeze(-1 ) lowerCamelCase__ : Union[str, Any] = -0.5 * beta_t * x lowerCamelCase__ : List[str] = torch.sqrt(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = drift - diffusion**2 * score lowerCamelCase__ : Dict = x + drift * dt # add noise lowerCamelCase__ : int = randn_tensor(x.shape , layout=x.layout , generator=UpperCamelCase__ , device=x.device , dtype=x.dtype ) lowerCamelCase__ : str = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self: Dict ): return self.config.num_train_timesteps
41
'''simple docstring''' from __future__ import annotations _A : Any ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class _lowercase : def __init__( self: Tuple , UpperCamelCase__: dict[str, list[str]] , UpperCamelCase__: str ): lowerCamelCase__ : str = graph # mapping node to its parent in resulting breadth first tree lowerCamelCase__ : dict[str, str | None] = {} lowerCamelCase__ : Any = source_vertex def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = {self.source_vertex} lowerCamelCase__ : List[str] = None lowerCamelCase__ : Tuple = [self.source_vertex] # first in first out queue while queue: lowerCamelCase__ : Tuple = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase__ ) lowerCamelCase__ : List[str] = vertex queue.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): if target_vertex == self.source_vertex: return self.source_vertex lowerCamelCase__ : Tuple = self.parent.get(UpperCamelCase__ ) if target_vertex_parent is None: lowerCamelCase__ : int = ( F'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(UpperCamelCase__ ) return self.shortest_path(UpperCamelCase__ ) + F'''->{target_vertex}''' if __name__ == "__main__": _A : int =Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
41
1
'''simple docstring''' from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lowercase ( _lowercase ): a = ["""image_processor""", """tokenizer"""] a = """BridgeTowerImageProcessor""" a = ("""RobertaTokenizer""", """RobertaTokenizerFast""") def __init__( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple ): super().__init__(UpperCamelCase__ , UpperCamelCase__ ) def __call__( self: Union[str, Any] , UpperCamelCase__: Any , UpperCamelCase__: Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , UpperCamelCase__: bool = True , UpperCamelCase__: Union[bool, str, PaddingStrategy] = False , UpperCamelCase__: Union[bool, str, TruncationStrategy] = None , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: int = 0 , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: Optional[bool] = None , UpperCamelCase__: Optional[bool] = None , UpperCamelCase__: bool = False , UpperCamelCase__: bool = False , UpperCamelCase__: bool = False , UpperCamelCase__: bool = False , UpperCamelCase__: bool = True , UpperCamelCase__: Optional[Union[str, TensorType]] = None , **UpperCamelCase__: Any , ): lowerCamelCase__ : int = self.tokenizer( text=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ , stride=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , return_overflowing_tokens=UpperCamelCase__ , return_special_tokens_mask=UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , return_length=UpperCamelCase__ , verbose=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ , ) # add pixel_values + pixel_mask lowerCamelCase__ : List[str] = self.image_processor( UpperCamelCase__ , return_tensors=UpperCamelCase__ , do_normalize=UpperCamelCase__ , do_center_crop=UpperCamelCase__ , **UpperCamelCase__ ) encoding.update(UpperCamelCase__ ) return encoding def lowerCamelCase_ ( self: List[Any] , *UpperCamelCase__: Dict , **UpperCamelCase__: Optional[int] ): return self.tokenizer.batch_decode(*UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple , *UpperCamelCase__: Tuple , **UpperCamelCase__: int ): return self.tokenizer.decode(*UpperCamelCase__ , **UpperCamelCase__ ) @property def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Dict = self.tokenizer.model_input_names lowerCamelCase__ : Optional[int] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
41
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : str = 0.0 for coeff in reversed(UpperCamelCase ): lowerCamelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": _A : Any =(0.0, 0.0, 5.0, 9.3, 7.0) _A : Optional[Any] =10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
41
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : List[Any] ={ '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class _lowercase ( _lowercase ): a = """vivit""" def __init__( self: List[str] , UpperCamelCase__: List[Any]=224 , UpperCamelCase__: Tuple=32 , UpperCamelCase__: int=[2, 16, 16] , UpperCamelCase__: Optional[Any]=3 , UpperCamelCase__: Dict=768 , UpperCamelCase__: Optional[int]=12 , UpperCamelCase__: Tuple=12 , UpperCamelCase__: List[str]=3_072 , UpperCamelCase__: Optional[int]="gelu_fast" , UpperCamelCase__: Union[str, Any]=0.0 , UpperCamelCase__: Any=0.0 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: Optional[Any]=1e-06 , UpperCamelCase__: List[str]=True , **UpperCamelCase__: List[Any] , ): lowerCamelCase__ : List[Any] = hidden_size lowerCamelCase__ : Dict = num_hidden_layers lowerCamelCase__ : List[str] = num_attention_heads lowerCamelCase__ : Any = intermediate_size lowerCamelCase__ : Union[str, Any] = hidden_act lowerCamelCase__ : Dict = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : Optional[int] = layer_norm_eps lowerCamelCase__ : int = image_size lowerCamelCase__ : str = num_frames lowerCamelCase__ : Optional[Any] = tubelet_size lowerCamelCase__ : int = num_channels lowerCamelCase__ : Any = qkv_bias super().__init__(**UpperCamelCase__ )
41
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _A : List[Any] ='''pt''' elif is_tf_available(): _A : Any ='''tf''' else: _A : List[str] ='''jax''' class _lowercase ( _lowercase , unittest.TestCase ): a = ByTaTokenizer a = False def lowerCamelCase_ ( self: str ): super().setUp() lowerCamelCase__ : str = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ByTaTokenizer.from_pretrained("""google/byt5-small""" ) def lowerCamelCase_ ( self: Any , **UpperCamelCase__: Tuple ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Any=False , UpperCamelCase__: Union[str, Any]=20 , UpperCamelCase__: Optional[int]=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase__ ) ): try: lowerCamelCase__ : Tuple = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Union[str, Any] = list(filter(lambda UpperCamelCase__ : re.match(R"""^[ a-zA-Z]+$""" , t[1] ) , UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = list(filter(lambda UpperCamelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase__ ) , UpperCamelCase__ ) ) if max_length is not None and len(UpperCamelCase__ ) > max_length: lowerCamelCase__ : Dict = toks[:max_length] if min_length is not None and len(UpperCamelCase__ ) < min_length and len(UpperCamelCase__ ) > 0: while len(UpperCamelCase__ ) < min_length: lowerCamelCase__ : Optional[Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : Union[str, Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : Union[str, Any] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) if " " not in output_txt and len(UpperCamelCase__ ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase__ ) ) if with_prefix_space: lowerCamelCase__ : str = """ """ + output_txt lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) return output_txt, output_ids def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : Union[str, Any] = tokenizer(["""hi</s>""", """I went to the gym</s>""", """</s>"""] ) lowerCamelCase__ : Optional[int] = tokenizer(["""hi""", """I went to the gym""", """"""] ) self.assertListEqual(batch_with_eos_added["""input_ids"""] , batch_without_eos_added["""input_ids"""] ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[Any] = self.ta_base_tokenizer lowerCamelCase__ : Dict = """Unicode €.""" lowerCamelCase__ : List[Any] = tokenizer(UpperCamelCase__ ) lowerCamelCase__ : List[str] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : Tuple = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """Unicode €.</s>""" ) lowerCamelCase__ : List[Any] = tokenizer("""e è é ê ë""" ) lowerCamelCase__ : List[Any] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : str = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """e è é ê ë</s>""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """e è é ê ë</s>""" ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : int = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off lowerCamelCase__ : List[str] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on lowerCamelCase__ : int = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) if FRAMEWORK != "jax": lowerCamelCase__ : Any = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : str = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase__ : Tuple = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase__ ) self.assertNotIn("""decoder_attention_mask""" , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = [ """Summary of the text.""", """Another summary.""", ] lowerCamelCase__ : Union[str, Any] = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.ta_base_tokenizer lowerCamelCase__ : str = ["""A long paragraph for summarization. </s>"""] lowerCamelCase__ : Optional[Any] = ["""Summary of the text. </s>"""] # fmt: off lowerCamelCase__ : Tuple = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] lowerCamelCase__ : Any = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ , text_target=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , batch["""input_ids"""][0] ) self.assertEqual(UpperCamelCase__ , batch["""labels"""][0] ) def lowerCamelCase_ ( self: Optional[int] ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : int = tempfile.mkdtemp() lowerCamelCase__ : List[str] = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Any = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) lowerCamelCase__ : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Any = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Union[str, Any] = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[Any] = json.load(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = [F'''<extra_id_{i}>''' for i in range(125 )] lowerCamelCase__ : int = added_tokens_extra_ids + [ """an_additional_special_token""" ] lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : Dict = tokenizer_class.from_pretrained( UpperCamelCase__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=UpperCamelCase__ )] lowerCamelCase__ : Any = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : str = tokenizer_class.from_pretrained(UpperCamelCase__ ) self.assertTrue(tokenizer.decode([255] ) == """""" ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: List[str] ): pass def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: int ): # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens lowerCamelCase__ : Dict = self.get_tokenizers(fast=UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Dict = ["""t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """x""", """t""", """</s>"""] lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : str = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] lowerCamelCase__ : str = 0 lowerCamelCase__ : Any = tokenizer.convert_ids_to_tokens( UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) for attr in attributes_list: setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [] ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [token_id_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [token_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [token_id_to_test_setters] )
41
1
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: Optional[Any]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: str=0.6 , UpperCamelCase__: str=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Optional[Any] = batch_size lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Optional[Any] = is_training lowerCamelCase__ : int = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : List[str] = mask_ratio lowerCamelCase__ : Optional[int] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Any = (image_size // patch_size) ** 2 lowerCamelCase__ : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches lowerCamelCase__ : List[str] = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Union[str, Any] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Optional[int] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = TFViTMAEModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : int = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : str = outputs_dict[0].numpy() lowerCamelCase__ : Optional[int] = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def lowerCamelCase_ ( self: Dict ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__: int ): lowerCamelCase__ : Optional[int] = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): lowerCamelCase__ : List[str] = v.numpy() else: lowerCamelCase__ : Union[str, Any] = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = prepare_numpy_arrays(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : Any = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: str ): # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase__ : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : Tuple = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[Any] = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } lowerCamelCase__ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase__ : List[str] = main_layer_class(UpperCamelCase__ ) lowerCamelCase__ : int = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase__ : List[str] = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) lowerCamelCase__ : Union[str, Any] = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : int = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) lowerCamelCase__ : int = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : Any = outputs.last_hidden_state.numpy() lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : int = outputs.logits.numpy() lowerCamelCase__ : Dict = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) lowerCamelCase__ : Dict = model_class.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : str = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase__ : Optional[Any] = 0 else: lowerCamelCase__ : Union[str, Any] = after_outputs["""logits"""].numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) def lowerCamelCase_ ( self: Any ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Optional[int] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) lowerCamelCase__ : int = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase__ : int = model_class.from_config(model.config ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @slow def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[Any] ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: List[str] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Tuple = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase__ : str = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Any = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1e-4 )
41
'''simple docstring''' from heapq import heappop, heappush import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> tuple[float | int, list[tuple[int, int]]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = grid.shape lowerCamelCase__ : List[str] = [-1, 1, 0, 0] lowerCamelCase__ : Dict = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] lowerCamelCase__ , lowerCamelCase__ : Any = [(0, source)], set() lowerCamelCase__ : Tuple = np.full((rows, cols) , np.inf ) lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[int] = np.empty((rows, cols) , dtype=UpperCamelCase ) lowerCamelCase__ : str = None while queue: ((lowerCamelCase__) , (lowerCamelCase__)) : List[str] = heappop(UpperCamelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: lowerCamelCase__ : Optional[int] = [] while (x, y) != source: path.append((x, y) ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = predecessors[x, y] path.append(UpperCamelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCamelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: lowerCamelCase__ : Any = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCamelCase , (dist + 1, (nx, ny)) ) lowerCamelCase__ : Union[str, Any] = dist + 1 lowerCamelCase__ : List[str] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin _A : List[str] =get_tests_dir('''fixtures/test_sentencepiece_bpe_char.model''') @require_sentencepiece @require_tokenizers class _lowercase ( _lowercase , unittest.TestCase ): a = SpeechTaTokenizer a = False a = True def lowerCamelCase_ ( self: List[Any] ): super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase__ : str = SpeechTaTokenizer(UpperCamelCase__ ) lowerCamelCase__ : Dict = AddedToken("""<mask>""" , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) lowerCamelCase__ : str = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase_ ( self: int , UpperCamelCase__: Optional[Any] ): lowerCamelCase__ : int = """this is a test""" lowerCamelCase__ : Union[str, Any] = """this is a test""" return input_text, output_text def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Tuple , UpperCamelCase__: Dict=False , UpperCamelCase__: Optional[Any]=20 , UpperCamelCase__: Any=5 ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.get_input_output_texts(UpperCamelCase__ ) lowerCamelCase__ : Tuple = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) return text, ids def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = """<pad>""" lowerCamelCase__ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase__ ) , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-4] , """œ""" ) self.assertEqual(vocab_keys[-2] , """<mask>""" ) self.assertEqual(vocab_keys[-1] , """<ctc_blank>""" ) self.assertEqual(len(UpperCamelCase__ ) , 81 ) def lowerCamelCase_ ( self: Union[str, Any] ): self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : int = self.get_tokenizers(do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : str = tokenizer.vocab_size lowerCamelCase__ : List[Any] = len(UpperCamelCase__ ) self.assertNotEqual(UpperCamelCase__ , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) lowerCamelCase__ : Dict = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] lowerCamelCase__ : List[str] = tokenizer.add_tokens(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.vocab_size lowerCamelCase__ : Tuple = len(UpperCamelCase__ ) self.assertNotEqual(UpperCamelCase__ , 0 ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , len(UpperCamelCase__ ) ) self.assertEqual(UpperCamelCase__ , all_size + len(UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=UpperCamelCase__ ) self.assertGreaterEqual(len(UpperCamelCase__ ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) lowerCamelCase__ : int = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} lowerCamelCase__ : Any = tokenizer.add_special_tokens(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.vocab_size lowerCamelCase__ : Dict = len(UpperCamelCase__ ) self.assertNotEqual(UpperCamelCase__ , 0 ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , len(UpperCamelCase__ ) ) self.assertEqual(UpperCamelCase__ , all_size_a + len(UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=UpperCamelCase__ ) self.assertGreaterEqual(len(UpperCamelCase__ ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = self.get_tokenizer() lowerCamelCase__ : Union[str, Any] = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(UpperCamelCase__ , [SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) lowerCamelCase__ : Tuple = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCamelCase__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) lowerCamelCase__ : Dict = tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) # fmt: off self.assertListEqual(UpperCamelCase__ , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on lowerCamelCase__ : Optional[int] = tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def lowerCamelCase_ ( self: List[str] ): # Use custom sequence because this tokenizer does not handle numbers. lowerCamelCase__ : Any = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off lowerCamelCase__ : Dict = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase__ , model_name="""microsoft/speecht5_asr""" , revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" , sequences=UpperCamelCase__ , )
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
'''simple docstring''' from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _lowercase ( _lowercase ): a = CustomTokenizer pass
41
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Any =logging.get_logger(__name__) _A : List[str] ={ '''edbeeching/decision-transformer-gym-hopper-medium''': ( '''https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json''' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class _lowercase ( _lowercase ): a = """decision_transformer""" a = ["""past_key_values"""] a = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self: Optional[Any] , UpperCamelCase__: Dict=17 , UpperCamelCase__: int=4 , UpperCamelCase__: Optional[int]=128 , UpperCamelCase__: int=4_096 , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[Any]=1 , UpperCamelCase__: Any=1_024 , UpperCamelCase__: Dict=3 , UpperCamelCase__: List[str]=1 , UpperCamelCase__: int=None , UpperCamelCase__: Union[str, Any]="relu" , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Dict=0.1 , UpperCamelCase__: int=0.1 , UpperCamelCase__: Optional[int]=1e-5 , UpperCamelCase__: str=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Any=True , UpperCamelCase__: int=50_256 , UpperCamelCase__: Union[str, Any]=50_256 , UpperCamelCase__: Optional[int]=False , UpperCamelCase__: Union[str, Any]=False , **UpperCamelCase__: List[Any] , ): lowerCamelCase__ : Optional[Any] = state_dim lowerCamelCase__ : Optional[int] = act_dim lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : str = max_ep_len lowerCamelCase__ : Optional[Any] = action_tanh lowerCamelCase__ : Dict = vocab_size lowerCamelCase__ : Dict = n_positions lowerCamelCase__ : Union[str, Any] = n_layer lowerCamelCase__ : List[Any] = n_head lowerCamelCase__ : int = n_inner lowerCamelCase__ : str = activation_function lowerCamelCase__ : str = resid_pdrop lowerCamelCase__ : Optional[Any] = embd_pdrop lowerCamelCase__ : Union[str, Any] = attn_pdrop lowerCamelCase__ : int = layer_norm_epsilon lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : str = scale_attn_weights lowerCamelCase__ : Tuple = use_cache lowerCamelCase__ : Any = scale_attn_by_inverse_layer_idx lowerCamelCase__ : Any = reorder_and_upcast_attn lowerCamelCase__ : List[Any] = bos_token_id lowerCamelCase__ : Dict = eos_token_id super().__init__(bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ )
41
'''simple docstring''' class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : dict[str, TrieNode] = {} # Mapping from char to TrieNode lowerCamelCase__ : List[str] = False def lowerCamelCase_ ( self: str , UpperCamelCase__: list[str] ): for word in words: self.insert(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : List[Any] = self for char in word: if char not in curr.nodes: lowerCamelCase__ : Tuple = TrieNode() lowerCamelCase__ : List[Any] = curr.nodes[char] lowerCamelCase__ : Any = True def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = self for char in word: if char not in curr.nodes: return False lowerCamelCase__ : Any = curr.nodes[char] return curr.is_leaf def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): def _delete(UpperCamelCase__: TrieNode , UpperCamelCase__: str , UpperCamelCase__: int ) -> bool: if index == len(UpperCamelCase__ ): # If word does not exist if not curr.is_leaf: return False lowerCamelCase__ : str = False return len(curr.nodes ) == 0 lowerCamelCase__ : List[str] = word[index] lowerCamelCase__ : Dict = curr.nodes.get(UpperCamelCase__ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCamelCase__ : List[Any] = _delete(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCamelCase__ , 0 ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: if node.is_leaf: print(UpperCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(UpperCamelCase , word + key ) def SCREAMING_SNAKE_CASE_ () -> bool: lowerCamelCase__ : str = """banana bananas bandana band apple all beast""".split() lowerCamelCase__ : Union[str, Any] = TrieNode() root.insert_many(UpperCamelCase ) # print_words(root, "") assert all(root.find(UpperCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: print(str(UpperCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def SCREAMING_SNAKE_CASE_ () -> None: assert test_trie() def SCREAMING_SNAKE_CASE_ () -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
41
1
'''simple docstring''' import os def SCREAMING_SNAKE_CASE_ () -> Optional[Any]: with open(os.path.dirname(UpperCamelCase ) + """/grid.txt""" ) as f: lowerCamelCase__ : Tuple = [] # noqa: E741 for _ in range(20 ): l.append([int(UpperCamelCase ) for x in f.readline().split()] ) lowerCamelCase__ : Dict = 0 # right for i in range(20 ): for j in range(17 ): lowerCamelCase__ : Any = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: lowerCamelCase__ : List[str] = temp # down for i in range(17 ): for j in range(20 ): lowerCamelCase__ : Dict = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: lowerCamelCase__ : Dict = temp # diagonal 1 for i in range(17 ): for j in range(17 ): lowerCamelCase__ : List[str] = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: lowerCamelCase__ : Optional[int] = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): lowerCamelCase__ : int = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: lowerCamelCase__ : Tuple = temp return maximum if __name__ == "__main__": print(solution())
41
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("""repo_id""" , ["""canonical_dataset_name""", """org-name/dataset-name"""] ) @pytest.mark.parametrize("""path""" , ["""filename.csv""", """filename with blanks.csv"""] ) @pytest.mark.parametrize("""revision""" , [None, """v2"""] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Dict = hf_hub_url(repo_id=UpperCamelCase , path=UpperCamelCase , revision=UpperCamelCase ) assert url == f'''https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(UpperCamelCase )}'''
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import os import string import sys _A : Optional[Any] =1 << 8 _A : Union[str, Any] ={ '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 27, '''up''': 65 + ARROW_KEY_FLAG, '''down''': 66 + ARROW_KEY_FLAG, '''right''': 67 + ARROW_KEY_FLAG, '''left''': 68 + ARROW_KEY_FLAG, '''mod_int''': 91, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 50, '''delete''': 51, '''pg_up''': 53, '''pg_down''': 54, } _A : Optional[Any] =KEYMAP['''up'''] _A : Optional[Any] =KEYMAP['''left'''] if sys.platform == "win32": _A : List[str] =[] _A : Union[str, Any] ={ b'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, b'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(10): _A : List[str] =ord(str(i)) def SCREAMING_SNAKE_CASE_ () -> int: if os.name == "nt": import msvcrt lowerCamelCase__ : str = """mbcs""" # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(UpperCamelCase ) == 0: # Read the keystroke lowerCamelCase__ : List[Any] = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): lowerCamelCase__ : Optional[int] = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: lowerCamelCase__ : Tuple = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP["""mod_int"""] ) ) WIN_CH_BUFFER.append(UpperCamelCase ) if ord(UpperCamelCase ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) lowerCamelCase__ : str = chr(KEYMAP["""esc"""] ) except KeyError: lowerCamelCase__ : List[str] = cha[1] else: lowerCamelCase__ : Any = ch.decode(UpperCamelCase ) else: lowerCamelCase__ : str = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty lowerCamelCase__ : List[Any] = sys.stdin.fileno() lowerCamelCase__ : Any = termios.tcgetattr(UpperCamelCase ) try: tty.setraw(UpperCamelCase ) lowerCamelCase__ : Dict = sys.stdin.read(1 ) finally: termios.tcsetattr(UpperCamelCase , termios.TCSADRAIN , UpperCamelCase ) return ch def SCREAMING_SNAKE_CASE_ () -> Optional[int]: lowerCamelCase__ : Dict = get_raw_chars() if ord(UpperCamelCase ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(UpperCamelCase ) == KEYMAP["esc"]: lowerCamelCase__ : List[str] = get_raw_chars() if ord(UpperCamelCase ) == KEYMAP["mod_int"]: lowerCamelCase__ : str = get_raw_chars() if ord(UpperCamelCase ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(UpperCamelCase ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(UpperCamelCase ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
41
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _A : List[Any] =logging.get_logger(__name__) _A : Dict =['''model.decoder.embed_positions.weights'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: if "emb" in name: lowerCamelCase__ : Dict = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: lowerCamelCase__ : List[str] = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: lowerCamelCase__ : List[str] = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: lowerCamelCase__ : Optional[int] = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: lowerCamelCase__ : Dict = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: lowerCamelCase__ : Optional[Any] = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: lowerCamelCase__ : Optional[Any] = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: lowerCamelCase__ : int = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple[Dict, Dict]: lowerCamelCase__ : int = list(state_dict.keys() ) lowerCamelCase__ : Tuple = {} for key in keys: lowerCamelCase__ : Any = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj lowerCamelCase__ : Union[str, Any] = val[:hidden_size, :] lowerCamelCase__ : Any = val[hidden_size : 2 * hidden_size, :] lowerCamelCase__ : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: lowerCamelCase__ : str = val else: lowerCamelCase__ : Union[str, Any] = val return state_dict, enc_dec_proj_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values lowerCamelCase__ : int = 1024 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 16 elif checkpoint == "medium": lowerCamelCase__ : Any = 1536 lowerCamelCase__ : Union[str, Any] = 48 lowerCamelCase__ : Optional[int] = 24 elif checkpoint == "large": lowerCamelCase__ : Optional[Any] = 2048 lowerCamelCase__ : Dict = 48 lowerCamelCase__ : List[Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) lowerCamelCase__ : Any = MusicgenDecoderConfig( hidden_size=UpperCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase , num_attention_heads=UpperCamelCase , ) return config @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase="cpu" ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = MusicGen.get_pretrained(UpperCamelCase , device=UpperCamelCase ) lowerCamelCase__ : List[Any] = decoder_config_from_checkpoint(UpperCamelCase ) lowerCamelCase__ : Any = fairseq_model.lm.state_dict() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = rename_state_dict( UpperCamelCase , hidden_size=decoder_config.hidden_size ) lowerCamelCase__ : str = TaEncoderModel.from_pretrained("""t5-base""" ) lowerCamelCase__ : Tuple = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) lowerCamelCase__ : Optional[int] = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection lowerCamelCase__ , lowerCamelCase__ : List[str] = decoder.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model lowerCamelCase__ : Optional[Any] = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase , audio_encoder=UpperCamelCase , decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass lowerCamelCase__ : Dict = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) lowerCamelCase__ : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor lowerCamelCase__ : str = AutoTokenizer.from_pretrained("""t5-base""" ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) lowerCamelCase__ : Optional[int] = MusicgenProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids lowerCamelCase__ : Union[str, Any] = 2048 lowerCamelCase__ : List[str] = 2048 # set other default generation config params lowerCamelCase__ : Optional[Any] = int(30 * audio_encoder.config.frame_rate ) lowerCamelCase__ : Union[str, Any] = True lowerCamelCase__ : List[Any] = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _A : List[str] =parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
41
1
'''simple docstring''' import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) _A : Optional[int] ='''hf-internal-testing/tiny-random-bert''' _A : Union[str, Any] =os.path.join(TRANSFORMERS_CACHE, '''models--hf-internal-testing--tiny-random-bert''') _A : Optional[Any] ='''9b8c223d42b2188cb49d29af482996f9d0f3e5a6''' class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Dict = cached_file(UpperCamelCase__ , UpperCamelCase__ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCamelCase__ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) ) with open(os.path.join(UpperCamelCase__ , """refs""" , """main""" ) ) as f: lowerCamelCase__ : int = f.read() self.assertEqual(UpperCamelCase__ , os.path.join(UpperCamelCase__ , """snapshots""" , UpperCamelCase__ , UpperCamelCase__ ) ) self.assertTrue(os.path.isfile(UpperCamelCase__ ) ) # File is cached at the same place the second time. lowerCamelCase__ : Union[str, Any] = cached_file(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) # Using a specific revision to test the full commit hash. lowerCamelCase__ : str = cached_file(UpperCamelCase__ , UpperCamelCase__ , revision="""9b8c223""" ) self.assertEqual(UpperCamelCase__ , os.path.join(UpperCamelCase__ , """snapshots""" , UpperCamelCase__ , UpperCamelCase__ ) ) def lowerCamelCase_ ( self: List[Any] ): with self.assertRaisesRegex(UpperCamelCase__ , """is not a valid model identifier""" ): lowerCamelCase__ : Tuple = cached_file("""tiny-random-bert""" , UpperCamelCase__ ) with self.assertRaisesRegex(UpperCamelCase__ , """is not a valid git identifier""" ): lowerCamelCase__ : List[str] = cached_file(UpperCamelCase__ , UpperCamelCase__ , revision="""aaaa""" ) with self.assertRaisesRegex(UpperCamelCase__ , """does not appear to have a file named""" ): lowerCamelCase__ : str = cached_file(UpperCamelCase__ , """conf""" ) def lowerCamelCase_ ( self: Optional[int] ): with self.assertRaisesRegex(UpperCamelCase__ , """does not appear to have a file named""" ): lowerCamelCase__ : Any = cached_file(UpperCamelCase__ , """conf""" ) with open(os.path.join(UpperCamelCase__ , """refs""" , """main""" ) ) as f: lowerCamelCase__ : Optional[int] = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase__ , """.no_exist""" , UpperCamelCase__ , """conf""" ) ) ) lowerCamelCase__ : Optional[Any] = cached_file(UpperCamelCase__ , """conf""" , _raise_exceptions_for_missing_entries=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) lowerCamelCase__ : Dict = cached_file(UpperCamelCase__ , """conf""" , local_files_only=UpperCamelCase__ , _raise_exceptions_for_missing_entries=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = mock.Mock() lowerCamelCase__ : str = 500 lowerCamelCase__ : List[str] = {} lowerCamelCase__ : Union[str, Any] = HTTPError lowerCamelCase__ : List[str] = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=UpperCamelCase__ ) as mock_head: lowerCamelCase__ : List[Any] = cached_file(UpperCamelCase__ , """conf""" , _raise_exceptions_for_connection_errors=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) # This check we did call the fake head request mock_head.assert_called() def lowerCamelCase_ ( self: Dict ): self.assertTrue(has_file("""hf-internal-testing/tiny-bert-pt-only""" , UpperCamelCase__ ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , UpperCamelCase__ ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , UpperCamelCase__ ) ) def lowerCamelCase_ ( self: Optional[Any] ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo("""bert-base-cased""" , """ahah.txt""" ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCamelCase__ , """is not a valid model identifier""" ): get_file_from_repo("""bert-base-case""" , UpperCamelCase__ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCamelCase__ , """is not a valid git identifier""" ): get_file_from_repo("""bert-base-cased""" , UpperCamelCase__ , revision="""ahaha""" ) lowerCamelCase__ : Tuple = get_file_from_repo("""bert-base-cased""" , UpperCamelCase__ ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCamelCase__ : str = json.loads(open(UpperCamelCase__ , """r""" ).read() ) self.assertEqual(config["""hidden_size"""] , 768 ) def lowerCamelCase_ ( self: List[Any] ): with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase__ : int = Path(UpperCamelCase__ ) / """a.txt""" filename.touch() self.assertEqual(get_file_from_repo(UpperCamelCase__ , """a.txt""" ) , str(UpperCamelCase__ ) ) self.assertIsNone(get_file_from_repo(UpperCamelCase__ , """b.txt""" ) )
41
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _A : Optional[int] =logging.get_logger(__name__) _A : List[str] ={ '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _A : int ={ '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _A : Dict ={ '''facebook/blenderbot_small-90M''': 512, } class _lowercase ( _lowercase ): a = VOCAB_FILES_NAMES a = PRETRAINED_VOCAB_FILES_MAP a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a = BlenderbotSmallTokenizer def __init__( self: Tuple , UpperCamelCase__: List[Any]=None , UpperCamelCase__: Dict=None , UpperCamelCase__: int="<|endoftext|>" , UpperCamelCase__: Optional[Any]="<|endoftext|>" , UpperCamelCase__: Tuple="<|endoftext|>" , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: Optional[Any]=True , **UpperCamelCase__: Union[str, Any] , ): super().__init__( ByteLevelBPETokenizer( vocab=UpperCamelCase__ , merges=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , ) , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase__ : str = add_prefix_space def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Optional[Any] , UpperCamelCase__: List[Any]=None ): lowerCamelCase__ : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self: int , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): lowerCamelCase__ : Any = [self.sep_token_id] lowerCamelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
1
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class _lowercase ( _lowercase ): def __init__( self: List[str] , *UpperCamelCase__: List[Any] , **UpperCamelCase__: List[Any] ): super().__init__(*UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Any = {} def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Optional[Any] , *UpperCamelCase__: Union[str, Any] , **UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : str = super().add_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) if num_added_tokens == 0: raise ValueError( F'''The tokenizer already contains the token {placeholder_token}. Please pass a different''' """ `placeholder_token` that is not already in the tokenizer.""" ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Optional[Any] , *UpperCamelCase__: List[Any] , UpperCamelCase__: Dict=1 , **UpperCamelCase__: int ): lowerCamelCase__ : Dict = [] if num_vec_per_token == 1: self.try_adding_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) output.append(UpperCamelCase__ ) else: lowerCamelCase__ : Any = [] for i in range(UpperCamelCase__ ): lowerCamelCase__ : Dict = placeholder_token + F'''_{i}''' self.try_adding_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) output.append(UpperCamelCase__ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F'''The tokenizer already has placeholder token {token} that can get confused with''' F''' {placeholder_token}keep placeholder tokens independent''' ) lowerCamelCase__ : Tuple = output def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: Any=1.0 ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : int = [] for i in range(len(UpperCamelCase__ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=UpperCamelCase__ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: lowerCamelCase__ : Optional[Any] = self.token_map[placeholder_token] lowerCamelCase__ : Tuple = tokens[: 1 + int(len(UpperCamelCase__ ) * prop_tokens_to_load )] if vector_shuffle: lowerCamelCase__ : List[str] = copy.copy(UpperCamelCase__ ) random.shuffle(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = text.replace(UpperCamelCase__ , """ """.join(UpperCamelCase__ ) ) return text def __call__( self: str , UpperCamelCase__: Optional[Any] , *UpperCamelCase__: List[Any] , UpperCamelCase__: str=False , UpperCamelCase__: Optional[int]=1.0 , **UpperCamelCase__: Optional[Any] ): return super().__call__( self.replace_placeholder_tokens_in_text( UpperCamelCase__ , vector_shuffle=UpperCamelCase__ , prop_tokens_to_load=UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ , ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Union[str, Any] , *UpperCamelCase__: str , UpperCamelCase__: Tuple=False , UpperCamelCase__: Tuple=1.0 , **UpperCamelCase__: List[Any] ): return super().encode( self.replace_placeholder_tokens_in_text( UpperCamelCase__ , vector_shuffle=UpperCamelCase__ , prop_tokens_to_load=UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ , )
41
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=7 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[str]=18 , UpperCamelCase__: Union[str, Any]=30 , UpperCamelCase__: List[str]=400 , UpperCamelCase__: Any=True , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[Any]=False , UpperCamelCase__: Tuple=True , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , ): lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Optional[int] = min_resolution lowerCamelCase__ : Optional[Any] = max_resolution lowerCamelCase__ : Union[str, Any] = do_resize lowerCamelCase__ : Optional[Any] = size if size is not None else {"""height""": 18, """width""": 20} lowerCamelCase__ : Dict = do_thumbnail lowerCamelCase__ : Optional[int] = do_align_axis lowerCamelCase__ : Any = do_pad lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Union[str, Any] = image_mean lowerCamelCase__ : Union[str, Any] = image_std def lowerCamelCase_ ( self: str ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DonutImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = DonutImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_thumbnail""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_align_long_axis""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 20} ) lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order lowerCamelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"""height""": 84, """width""": 42} ) def lowerCamelCase_ ( self: List[str] ): pass @is_flaky() def lowerCamelCase_ ( self: Union[str, Any] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Optional[int] ): # Initialize image_processing lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Dict ): # Initialize image_processing lowerCamelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
41
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A : Tuple =logging.get_logger(__name__) _A : str ={ '''google/mobilenet_v2_1.4_224''': '''https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json''', '''google/mobilenet_v2_1.0_224''': '''https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json''', '''google/mobilenet_v2_0.75_160''': '''https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json''', '''google/mobilenet_v2_0.35_96''': '''https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json''', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class _lowercase ( _lowercase ): a = """mobilenet_v2""" def __init__( self: List[Any] , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: Dict=224 , UpperCamelCase__: Optional[Any]=1.0 , UpperCamelCase__: Tuple=8 , UpperCamelCase__: Dict=8 , UpperCamelCase__: List[str]=6 , UpperCamelCase__: Tuple=32 , UpperCamelCase__: List[Any]=True , UpperCamelCase__: str=True , UpperCamelCase__: Dict="relu6" , UpperCamelCase__: Dict=True , UpperCamelCase__: Union[str, Any]=0.8 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: str=0.001 , UpperCamelCase__: Union[str, Any]=255 , **UpperCamelCase__: Dict , ): super().__init__(**UpperCamelCase__ ) if depth_multiplier <= 0: raise ValueError("""depth_multiplier must be greater than zero.""" ) lowerCamelCase__ : str = num_channels lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : str = depth_multiplier lowerCamelCase__ : Optional[Any] = depth_divisible_by lowerCamelCase__ : List[str] = min_depth lowerCamelCase__ : Tuple = expand_ratio lowerCamelCase__ : Union[str, Any] = output_stride lowerCamelCase__ : str = first_layer_is_expansion lowerCamelCase__ : List[Any] = finegrained_output lowerCamelCase__ : Tuple = hidden_act lowerCamelCase__ : int = tf_padding lowerCamelCase__ : List[str] = classifier_dropout_prob lowerCamelCase__ : Optional[Any] = initializer_range lowerCamelCase__ : Union[str, Any] = layer_norm_eps lowerCamelCase__ : List[str] = semantic_loss_ignore_index class _lowercase ( _lowercase ): a = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self: Tuple ): return OrderedDict([("""pixel_values""", {0: """batch"""})] ) @property def lowerCamelCase_ ( self: Optional[Any] ): if self.task == "image-classification": return OrderedDict([("""logits""", {0: """batch"""})] ) else: return OrderedDict([("""last_hidden_state""", {0: """batch"""}), ("""pooler_output""", {0: """batch"""})] ) @property def lowerCamelCase_ ( self: str ): return 1e-4
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> list[str]: if nth_term == "": return [""] lowerCamelCase__ : Any = int(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = int(UpperCamelCase ) lowerCamelCase__ : list[str] = [] for temp in range(int(UpperCamelCase ) ): series.append(f'''1 / {pow(temp + 1 , int(UpperCamelCase ) )}''' if series else """1""" ) return series if __name__ == "__main__": import doctest doctest.testmod() _A : List[str] =int(input('''Enter the last number (nth term) of the P-Series''')) _A : int =int(input('''Enter the power for P-Series''')) print('''Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p''') print(p_series(nth_term, power))
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _A : str ='''pt''' elif is_tf_available(): _A : Optional[Any] ='''tf''' else: _A : Tuple ='''jax''' class _lowercase ( _lowercase , unittest.TestCase ): a = PerceiverTokenizer a = False def lowerCamelCase_ ( self: str ): super().setUp() lowerCamelCase__ : Optional[int] = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Optional[int] ): return PerceiverTokenizer.from_pretrained("""deepmind/language-perceiver""" ) def lowerCamelCase_ ( self: Optional[int] , **UpperCamelCase__: str ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Tuple , UpperCamelCase__: List[str]=False , UpperCamelCase__: Any=20 , UpperCamelCase__: str=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase__ ) ): try: lowerCamelCase__ : str = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Any = list(filter(lambda UpperCamelCase__ : re.match(R"""^[ a-zA-Z]+$""" , t[1] ) , UpperCamelCase__ ) ) lowerCamelCase__ : str = list(filter(lambda UpperCamelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase__ ) , UpperCamelCase__ ) ) if max_length is not None and len(UpperCamelCase__ ) > max_length: lowerCamelCase__ : str = toks[:max_length] if min_length is not None and len(UpperCamelCase__ ) < min_length and len(UpperCamelCase__ ) > 0: while len(UpperCamelCase__ ) < min_length: lowerCamelCase__ : Any = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : Optional[Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : List[Any] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) if " " not in output_txt and len(UpperCamelCase__ ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase__ ) ) if with_prefix_space: lowerCamelCase__ : List[Any] = """ """ + output_txt lowerCamelCase__ : str = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) return output_txt, output_ids def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Union[str, Any] = self.perceiver_tokenizer lowerCamelCase__ : List[str] = """Unicode €.""" lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) lowerCamelCase__ : str = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : Tuple = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """[CLS]Unicode €.[SEP]""" ) lowerCamelCase__ : Dict = tokenizer("""e è é ê ë""" ) lowerCamelCase__ : str = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : List[str] = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """[CLS]e è é ê ë[SEP]""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """[CLS]e è é ê ë[SEP]""" ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = self.perceiver_tokenizer lowerCamelCase__ : List[Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off lowerCamelCase__ : Optional[Any] = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on lowerCamelCase__ : Optional[Any] = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) if FRAMEWORK != "jax": lowerCamelCase__ : Union[str, Any] = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : Dict = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.perceiver_tokenizer lowerCamelCase__ : int = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase__ : Optional[int] = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase__ ) self.assertNotIn("""decoder_attention_mask""" , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = self.perceiver_tokenizer lowerCamelCase__ : Dict = [ """Summary of the text.""", """Another summary.""", ] lowerCamelCase__ : Any = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: str ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Optional[int] = tempfile.mkdtemp() lowerCamelCase__ : int = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : Any = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Optional[int] = tempfile.mkdtemp() lowerCamelCase__ : str = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) lowerCamelCase__ : Any = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : int = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Optional[int] = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Tuple = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[Any] = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Dict = json.load(UpperCamelCase__ ) lowerCamelCase__ : int = [F'''<extra_id_{i}>''' for i in range(125 )] lowerCamelCase__ : str = added_tokens_extra_ids + [ """an_additional_special_token""" ] lowerCamelCase__ : Any = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : Union[str, Any] = tokenizer_class.from_pretrained( UpperCamelCase__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : int = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=UpperCamelCase__ )] lowerCamelCase__ : Any = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Union[str, Any] = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) , """�""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass def lowerCamelCase_ ( self: Optional[Any] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: List[Any] ): pass def lowerCamelCase_ ( self: Union[str, Any] ): # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens lowerCamelCase__ : List[Any] = self.get_tokenizers(fast=UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : List[str] = ["""[CLS]""", """t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """s""", """t""", """[SEP]"""] lowerCamelCase__ : Tuple = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ )
41
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowerCamelCase__ : List[Any] = torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = pipe( image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowerCamelCase__ : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase__ : Tuple = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if not isinstance(UpperCamelCase , UpperCamelCase ): lowerCamelCase__ : Tuple = f'''Input value of [number={number}] must be an integer''' raise TypeError(UpperCamelCase ) if number < 1: lowerCamelCase__ : int = f'''Input value of [number={number}] must be > 0''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : Optional[int] = 1 for i in range(1 , UpperCamelCase ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _A : Dict ='''tiny-wmt19-en-ru''' # Build # borrowed from a test _A : List[str] =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _A : str =dict(zip(vocab, range(len(vocab)))) _A : List[str] =['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: _A : Union[str, Any] =Path(tmpdirname) _A : str =build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] _A : int =build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] _A : List[Any] =build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) _A : int =FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _A : List[str] =FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _A : Union[str, Any] =FSMTForConditionalGeneration(config) print(F'num of params {tiny_model.num_parameters()}') # Test _A : List[str] =tokenizer(['''Making tiny model'''], return_tensors='''pt''') _A : Tuple =tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'Generated {mname_tiny}') # Upload # transformers-cli upload tiny-wmt19-en-ru
41
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[int] =logging.get_logger(__name__) _A : Optional[Any] ={ '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/config.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/config.json''' # See all FNet models at https://huggingface.co/models?filter=fnet } class _lowercase ( _lowercase ): a = """fnet""" def __init__( self: Dict , UpperCamelCase__: List[Any]=32_000 , UpperCamelCase__: Optional[int]=768 , UpperCamelCase__: List[str]=12 , UpperCamelCase__: str=3_072 , UpperCamelCase__: List[Any]="gelu_new" , UpperCamelCase__: Optional[Any]=0.1 , UpperCamelCase__: List[str]=512 , UpperCamelCase__: List[Any]=4 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: Union[str, Any]=1e-12 , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: Optional[int]=512 , UpperCamelCase__: Union[str, Any]=3 , UpperCamelCase__: Optional[Any]=1 , UpperCamelCase__: Optional[int]=2 , **UpperCamelCase__: Dict , ): super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : List[Any] = vocab_size lowerCamelCase__ : Optional[Any] = max_position_embeddings lowerCamelCase__ : List[Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : str = intermediate_size lowerCamelCase__ : Dict = hidden_act lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : List[Any] = initializer_range lowerCamelCase__ : Optional[int] = type_vocab_size lowerCamelCase__ : Tuple = layer_norm_eps lowerCamelCase__ : Optional[int] = use_tpu_fourier_optimizations lowerCamelCase__ : int = tpu_short_seq_length
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' import random from typing import Any def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> list[Any]: for _ in range(len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = random.randint(0 , len(UpperCamelCase ) - 1 ) lowerCamelCase__ : Tuple = random.randint(0 , len(UpperCamelCase ) - 1 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = data[b], data[a] return data if __name__ == "__main__": _A : int =[0, 1, 2, 3, 4, 5, 6, 7] _A : Optional[Any] =['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
41
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: Optional[Any]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: str=0.6 , UpperCamelCase__: str=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Optional[Any] = batch_size lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Optional[Any] = is_training lowerCamelCase__ : int = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : List[str] = mask_ratio lowerCamelCase__ : Optional[int] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Any = (image_size // patch_size) ** 2 lowerCamelCase__ : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches lowerCamelCase__ : List[str] = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Union[str, Any] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Optional[int] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = TFViTMAEModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : int = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : str = outputs_dict[0].numpy() lowerCamelCase__ : Optional[int] = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def lowerCamelCase_ ( self: Dict ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__: int ): lowerCamelCase__ : Optional[int] = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): lowerCamelCase__ : List[str] = v.numpy() else: lowerCamelCase__ : Union[str, Any] = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = prepare_numpy_arrays(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : Any = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: str ): # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase__ : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : Tuple = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[Any] = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } lowerCamelCase__ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase__ : List[str] = main_layer_class(UpperCamelCase__ ) lowerCamelCase__ : int = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase__ : List[str] = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) lowerCamelCase__ : Union[str, Any] = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : int = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) lowerCamelCase__ : int = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : Any = outputs.last_hidden_state.numpy() lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : int = outputs.logits.numpy() lowerCamelCase__ : Dict = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) lowerCamelCase__ : Dict = model_class.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : str = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase__ : Optional[Any] = 0 else: lowerCamelCase__ : Union[str, Any] = after_outputs["""logits"""].numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) def lowerCamelCase_ ( self: Any ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Optional[int] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) lowerCamelCase__ : int = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase__ : int = model_class.from_config(model.config ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @slow def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[Any] ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: List[str] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Tuple = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase__ : str = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Any = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1e-4 )
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) _A : Dict ={'''configuration_beit''': ['''BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BeitConfig''', '''BeitOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Union[str, Any] =['''BeitFeatureExtractor'''] _A : Union[str, Any] =['''BeitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : str =[ '''BEIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BeitForImageClassification''', '''BeitForMaskedImageModeling''', '''BeitForSemanticSegmentation''', '''BeitModel''', '''BeitPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Tuple =[ '''FlaxBeitForImageClassification''', '''FlaxBeitForMaskedImageModeling''', '''FlaxBeitModel''', '''FlaxBeitPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' class _lowercase : def __init__( self: Tuple , UpperCamelCase__: list[int] ): lowerCamelCase__ : Union[str, Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [0] * len_array if len_array > 0: lowerCamelCase__ : Union[str, Any] = array[0] for i in range(1 , UpperCamelCase__ ): lowerCamelCase__ : Tuple = self.prefix_sum[i - 1] + array[i] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: int ): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : Dict = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(UpperCamelCase__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Union[str, Any] ={ '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowercase ( _lowercase ): a = """roformer""" def __init__( self: List[Any] , UpperCamelCase__: Any=50_000 , UpperCamelCase__: List[Any]=None , UpperCamelCase__: Optional[int]=768 , UpperCamelCase__: Tuple=12 , UpperCamelCase__: Union[str, Any]=12 , UpperCamelCase__: Optional[Any]=3_072 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Any=0.1 , UpperCamelCase__: List[str]=0.1 , UpperCamelCase__: int=1_536 , UpperCamelCase__: int=2 , UpperCamelCase__: Dict=0.02 , UpperCamelCase__: Optional[int]=1e-12 , UpperCamelCase__: Optional[int]=0 , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: Optional[Any]=True , **UpperCamelCase__: List[Any] , ): super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Dict = vocab_size lowerCamelCase__ : List[Any] = hidden_size if embedding_size is None else embedding_size lowerCamelCase__ : Dict = hidden_size lowerCamelCase__ : Tuple = num_hidden_layers lowerCamelCase__ : List[Any] = num_attention_heads lowerCamelCase__ : List[Any] = hidden_act lowerCamelCase__ : Union[str, Any] = intermediate_size lowerCamelCase__ : str = hidden_dropout_prob lowerCamelCase__ : List[Any] = attention_probs_dropout_prob lowerCamelCase__ : Any = max_position_embeddings lowerCamelCase__ : Dict = type_vocab_size lowerCamelCase__ : List[str] = initializer_range lowerCamelCase__ : str = layer_norm_eps lowerCamelCase__ : Optional[Any] = rotary_value lowerCamelCase__ : Union[str, Any] = use_cache class _lowercase ( _lowercase ): @property def lowerCamelCase_ ( self: Tuple ): if self.task == "multiple-choice": lowerCamelCase__ : str = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase__ : int = {0: """batch""", 1: """sequence"""} lowerCamelCase__ : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A : Dict ={ '''configuration_swinv2''': ['''SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Swinv2Config'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Swinv2ForImageClassification''', '''Swinv2ForMaskedImageModeling''', '''Swinv2Model''', '''Swinv2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: Any=13 , UpperCamelCase__: List[str]=30 , UpperCamelCase__: Optional[Any]=2 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: int=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: List[Any]=4 , UpperCamelCase__: str=37 , UpperCamelCase__: str="gelu" , UpperCamelCase__: Tuple=0.1 , UpperCamelCase__: Optional[Any]=0.1 , UpperCamelCase__: Optional[int]=10 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Dict=3 , UpperCamelCase__: Optional[int]=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : str = batch_size lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : Tuple = patch_size lowerCamelCase__ : Dict = num_channels lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : Optional[int] = use_labels lowerCamelCase__ : Optional[int] = hidden_size lowerCamelCase__ : Union[str, Any] = num_hidden_layers lowerCamelCase__ : int = num_attention_heads lowerCamelCase__ : Union[str, Any] = intermediate_size lowerCamelCase__ : Union[str, Any] = hidden_act lowerCamelCase__ : Union[str, Any] = hidden_dropout_prob lowerCamelCase__ : Optional[int] = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : int = initializer_range lowerCamelCase__ : Optional[Any] = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase__ : str = (image_size // patch_size) ** 2 lowerCamelCase__ : Optional[Any] = num_patches + 1 def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : int = None if self.use_labels: lowerCamelCase__ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[int] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Union[str, Any] ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Any , UpperCamelCase__: Dict ): lowerCamelCase__ : str = TFViTModel(config=UpperCamelCase__ ) lowerCamelCase__ : Dict = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase__ : Tuple = self.image_size // 2 lowerCamelCase__ : Dict = pixel_values[:, :, :image_size, :image_size] lowerCamelCase__ : Dict = model(UpperCamelCase__ , interpolate_pos_encoding=UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Tuple ): lowerCamelCase__ : List[str] = self.type_sequence_label_size lowerCamelCase__ : Optional[Any] = TFViTForImageClassification(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , labels=UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. lowerCamelCase__ : str = self.image_size // 2 lowerCamelCase__ : str = pixel_values[:, :, :image_size, :image_size] lowerCamelCase__ : str = model(UpperCamelCase__ , interpolate_pos_encoding=UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase__ : Optional[Any] = 1 lowerCamelCase__ : Tuple = TFViTForImageClassification(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = config_and_inputs lowerCamelCase__ : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () a = ( {"""feature-extraction""": TFViTModel, """image-classification""": TFViTForImageClassification} if is_tf_available() else {} ) a = False a = False a = False def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = TFViTModelTester(self ) lowerCamelCase__ : Any = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Optional[Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def lowerCamelCase_ ( self: int ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[Any] = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[str] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): lowerCamelCase__ : List[str] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Optional[Any]: lowerCamelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Any ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) lowerCamelCase__ : str = self.default_image_processor lowerCamelCase__ : Optional[Any] = prepare_img() lowerCamelCase__ : Optional[int] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # forward pass lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : int = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : Dict = tf.constant([-0.2_744, 0.8_215, -0.0_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 )
41
'''simple docstring''' from __future__ import annotations _A : Any ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class _lowercase : def __init__( self: Tuple , UpperCamelCase__: dict[str, list[str]] , UpperCamelCase__: str ): lowerCamelCase__ : str = graph # mapping node to its parent in resulting breadth first tree lowerCamelCase__ : dict[str, str | None] = {} lowerCamelCase__ : Any = source_vertex def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = {self.source_vertex} lowerCamelCase__ : List[str] = None lowerCamelCase__ : Tuple = [self.source_vertex] # first in first out queue while queue: lowerCamelCase__ : Tuple = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase__ ) lowerCamelCase__ : List[str] = vertex queue.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): if target_vertex == self.source_vertex: return self.source_vertex lowerCamelCase__ : Tuple = self.parent.get(UpperCamelCase__ ) if target_vertex_parent is None: lowerCamelCase__ : int = ( F'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(UpperCamelCase__ ) return self.shortest_path(UpperCamelCase__ ) + F'''->{target_vertex}''' if __name__ == "__main__": _A : int =Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
41
1
'''simple docstring''' import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class _lowercase ( unittest.TestCase , _lowercase ): def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Union[str, Any] = load_tool("""text-classification""" ) self.tool.setup() lowerCamelCase__ : Optional[int] = load_tool("""text-classification""" , remote=UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[int] = self.tool("""That's quite cool""" , ["""positive""", """negative"""] ) self.assertEqual(UpperCamelCase__ , """positive""" ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[int] = self.remote_tool("""That's quite cool""" , ["""positive""", """negative"""] ) self.assertEqual(UpperCamelCase__ , """positive""" ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.tool(text="""That's quite cool""" , labels=["""positive""", """negative"""] ) self.assertEqual(UpperCamelCase__ , """positive""" ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : List[Any] = self.remote_tool(text="""That's quite cool""" , labels=["""positive""", """negative"""] ) self.assertEqual(UpperCamelCase__ , """positive""" )
41
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : str = 0.0 for coeff in reversed(UpperCamelCase ): lowerCamelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": _A : Any =(0.0, 0.0, 5.0, 9.3, 7.0) _A : Optional[Any] =10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _A : int ={ '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Tuple =['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[str] =[ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys _A : str =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _A : List[Any] ='''pt''' elif is_tf_available(): _A : Any ='''tf''' else: _A : List[str] ='''jax''' class _lowercase ( _lowercase , unittest.TestCase ): a = ByTaTokenizer a = False def lowerCamelCase_ ( self: str ): super().setUp() lowerCamelCase__ : str = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ByTaTokenizer.from_pretrained("""google/byt5-small""" ) def lowerCamelCase_ ( self: Any , **UpperCamelCase__: Tuple ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Any=False , UpperCamelCase__: Union[str, Any]=20 , UpperCamelCase__: Optional[int]=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase__ ) ): try: lowerCamelCase__ : Tuple = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Union[str, Any] = list(filter(lambda UpperCamelCase__ : re.match(R"""^[ a-zA-Z]+$""" , t[1] ) , UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = list(filter(lambda UpperCamelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase__ ) , UpperCamelCase__ ) ) if max_length is not None and len(UpperCamelCase__ ) > max_length: lowerCamelCase__ : Dict = toks[:max_length] if min_length is not None and len(UpperCamelCase__ ) < min_length and len(UpperCamelCase__ ) > 0: while len(UpperCamelCase__ ) < min_length: lowerCamelCase__ : Optional[Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : Union[str, Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : Union[str, Any] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) if " " not in output_txt and len(UpperCamelCase__ ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase__ ) ) if with_prefix_space: lowerCamelCase__ : str = """ """ + output_txt lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) return output_txt, output_ids def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : Union[str, Any] = tokenizer(["""hi</s>""", """I went to the gym</s>""", """</s>"""] ) lowerCamelCase__ : Optional[int] = tokenizer(["""hi""", """I went to the gym""", """"""] ) self.assertListEqual(batch_with_eos_added["""input_ids"""] , batch_without_eos_added["""input_ids"""] ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[Any] = self.ta_base_tokenizer lowerCamelCase__ : Dict = """Unicode €.""" lowerCamelCase__ : List[Any] = tokenizer(UpperCamelCase__ ) lowerCamelCase__ : List[str] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : Tuple = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """Unicode €.</s>""" ) lowerCamelCase__ : List[Any] = tokenizer("""e è é ê ë""" ) lowerCamelCase__ : List[Any] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : str = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """e è é ê ë</s>""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """e è é ê ë</s>""" ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : int = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off lowerCamelCase__ : List[str] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on lowerCamelCase__ : int = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) if FRAMEWORK != "jax": lowerCamelCase__ : Any = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : str = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase__ : Tuple = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase__ ) self.assertNotIn("""decoder_attention_mask""" , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = [ """Summary of the text.""", """Another summary.""", ] lowerCamelCase__ : Union[str, Any] = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.ta_base_tokenizer lowerCamelCase__ : str = ["""A long paragraph for summarization. </s>"""] lowerCamelCase__ : Optional[Any] = ["""Summary of the text. </s>"""] # fmt: off lowerCamelCase__ : Tuple = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] lowerCamelCase__ : Any = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ , text_target=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , batch["""input_ids"""][0] ) self.assertEqual(UpperCamelCase__ , batch["""labels"""][0] ) def lowerCamelCase_ ( self: Optional[int] ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : int = tempfile.mkdtemp() lowerCamelCase__ : List[str] = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Any = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) lowerCamelCase__ : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Any = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Union[str, Any] = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[Any] = json.load(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = [F'''<extra_id_{i}>''' for i in range(125 )] lowerCamelCase__ : int = added_tokens_extra_ids + [ """an_additional_special_token""" ] lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : Dict = tokenizer_class.from_pretrained( UpperCamelCase__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=UpperCamelCase__ )] lowerCamelCase__ : Any = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : str = tokenizer_class.from_pretrained(UpperCamelCase__ ) self.assertTrue(tokenizer.decode([255] ) == """""" ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: List[str] ): pass def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: int ): # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens lowerCamelCase__ : Dict = self.get_tokenizers(fast=UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Dict = ["""t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """x""", """t""", """</s>"""] lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : str = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] lowerCamelCase__ : str = 0 lowerCamelCase__ : Any = tokenizer.convert_ids_to_tokens( UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) for attr in attributes_list: setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [] ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [token_id_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [token_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [token_id_to_test_setters] )
41
1
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any]=13 , UpperCamelCase__: Optional[Any]=7 , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[str]=True , UpperCamelCase__: str=True , UpperCamelCase__: Optional[int]=True , UpperCamelCase__: Tuple=99 , UpperCamelCase__: Dict=32 , UpperCamelCase__: str=5 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: str=37 , UpperCamelCase__: str="gelu" , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: List[str]=512 , UpperCamelCase__: List[Any]=16 , UpperCamelCase__: str=2 , UpperCamelCase__: Union[str, Any]=0.02 , UpperCamelCase__: Any=4 , ): lowerCamelCase__ : Any = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Union[str, Any] = seq_length lowerCamelCase__ : List[str] = is_training lowerCamelCase__ : Optional[Any] = use_attention_mask lowerCamelCase__ : Any = use_token_type_ids lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : List[str] = vocab_size lowerCamelCase__ : Dict = hidden_size lowerCamelCase__ : Any = num_hidden_layers lowerCamelCase__ : List[Any] = num_attention_heads lowerCamelCase__ : List[Any] = intermediate_size lowerCamelCase__ : str = hidden_act lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : int = attention_probs_dropout_prob lowerCamelCase__ : List[Any] = max_position_embeddings lowerCamelCase__ : Union[str, Any] = type_vocab_size lowerCamelCase__ : str = type_sequence_label_size lowerCamelCase__ : int = initializer_range lowerCamelCase__ : List[Any] = num_choices def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : int = None if self.use_attention_mask: lowerCamelCase__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase__ : str = None if self.use_token_type_ids: lowerCamelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase__ : List[Any] = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCamelCase_ ( self: str ): lowerCamelCase__ : int = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = config_and_inputs lowerCamelCase__ : Tuple = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class _lowercase ( _lowercase , unittest.TestCase ): a = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : int = FlaxAlbertModelTester(self ) @slow def lowerCamelCase_ ( self: str ): for model_class_name in self.all_model_classes: lowerCamelCase__ : Union[str, Any] = model_class_name.from_pretrained("""albert-base-v2""" ) lowerCamelCase__ : Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class _lowercase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : Dict = FlaxAlbertModel.from_pretrained("""albert-base-v2""" ) lowerCamelCase__ : Dict = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) lowerCamelCase__ : List[str] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowerCamelCase__ : int = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ )[0] lowerCamelCase__ : List[Any] = (1, 11, 768) self.assertEqual(output.shape , UpperCamelCase__ ) lowerCamelCase__ : int = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCamelCase__ , atol=1e-4 ) )
41
'''simple docstring''' from heapq import heappop, heappush import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> tuple[float | int, list[tuple[int, int]]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = grid.shape lowerCamelCase__ : List[str] = [-1, 1, 0, 0] lowerCamelCase__ : Dict = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] lowerCamelCase__ , lowerCamelCase__ : Any = [(0, source)], set() lowerCamelCase__ : Tuple = np.full((rows, cols) , np.inf ) lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[int] = np.empty((rows, cols) , dtype=UpperCamelCase ) lowerCamelCase__ : str = None while queue: ((lowerCamelCase__) , (lowerCamelCase__)) : List[str] = heappop(UpperCamelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: lowerCamelCase__ : Optional[int] = [] while (x, y) != source: path.append((x, y) ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = predecessors[x, y] path.append(UpperCamelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCamelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: lowerCamelCase__ : Any = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCamelCase , (dist + 1, (nx, ny)) ) lowerCamelCase__ : Union[str, Any] = dist + 1 lowerCamelCase__ : List[str] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("""multiplicative_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""multiplicative_persistence() does not accept negative values""" ) lowerCamelCase__ : Union[str, Any] = 0 lowerCamelCase__ : List[str] = str(UpperCamelCase ) while len(UpperCamelCase ) != 1: lowerCamelCase__ : Tuple = [int(UpperCamelCase ) for i in num_string] lowerCamelCase__ : Any = 1 for i in range(0 , len(UpperCamelCase ) ): total *= numbers[i] lowerCamelCase__ : int = str(UpperCamelCase ) steps += 1 return steps def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("""additive_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""additive_persistence() does not accept negative values""" ) lowerCamelCase__ : str = 0 lowerCamelCase__ : Optional[int] = str(UpperCamelCase ) while len(UpperCamelCase ) != 1: lowerCamelCase__ : Tuple = [int(UpperCamelCase ) for i in num_string] lowerCamelCase__ : Any = 0 for i in range(0 , len(UpperCamelCase ) ): total += numbers[i] lowerCamelCase__ : List[Any] = str(UpperCamelCase ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
'''simple docstring''' _A : List[str] ='''Alexander Joslin''' import operator as op from .stack import Stack def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: lowerCamelCase__ : Dict = {"""*""": op.mul, """/""": op.truediv, """+""": op.add, """-""": op.sub} lowerCamelCase__ : Stack[int] = Stack() lowerCamelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(UpperCamelCase ) ) elif i in operators: # RULE 2 operator_stack.push(UpperCamelCase ) elif i == ")": # RULE 4 lowerCamelCase__ : Optional[Any] = operator_stack.peek() operator_stack.pop() lowerCamelCase__ : Dict = operand_stack.peek() operand_stack.pop() lowerCamelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCamelCase__ : Optional[int] = operators[opr](UpperCamelCase , UpperCamelCase ) operand_stack.push(UpperCamelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": _A : Optional[Any] ='''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F'{equation} = {dijkstras_two_stack_algorithm(equation)}')
41
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _A : str =logging.get_logger(__name__) class _lowercase ( _lowercase ): def __init__( self: List[Any] , *UpperCamelCase__: Union[str, Any] , **UpperCamelCase__: List[Any] ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , UpperCamelCase__ , ) super().__init__(*UpperCamelCase__ , **UpperCamelCase__ )
41
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _A : Dict ={ '''configuration_clip''': [ '''CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPConfig''', '''CLIPOnnxConfig''', '''CLIPTextConfig''', '''CLIPVisionConfig''', ], '''processing_clip''': ['''CLIPProcessor'''], '''tokenization_clip''': ['''CLIPTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Dict =['''CLIPTokenizerFast'''] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''CLIPFeatureExtractor'''] _A : List[Any] =['''CLIPImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[str] =[ '''CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPModel''', '''CLIPPreTrainedModel''', '''CLIPTextModel''', '''CLIPTextModelWithProjection''', '''CLIPVisionModel''', '''CLIPVisionModelWithProjection''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Union[str, Any] =[ '''TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFCLIPModel''', '''TFCLIPPreTrainedModel''', '''TFCLIPTextModel''', '''TFCLIPVisionModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Dict =[ '''FlaxCLIPModel''', '''FlaxCLIPPreTrainedModel''', '''FlaxCLIPTextModel''', '''FlaxCLIPTextPreTrainedModel''', '''FlaxCLIPVisionModel''', '''FlaxCLIPVisionPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : dict[str, TrieNode] = {} # Mapping from char to TrieNode lowerCamelCase__ : List[str] = False def lowerCamelCase_ ( self: str , UpperCamelCase__: list[str] ): for word in words: self.insert(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : List[Any] = self for char in word: if char not in curr.nodes: lowerCamelCase__ : Tuple = TrieNode() lowerCamelCase__ : List[Any] = curr.nodes[char] lowerCamelCase__ : Any = True def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = self for char in word: if char not in curr.nodes: return False lowerCamelCase__ : Any = curr.nodes[char] return curr.is_leaf def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): def _delete(UpperCamelCase__: TrieNode , UpperCamelCase__: str , UpperCamelCase__: int ) -> bool: if index == len(UpperCamelCase__ ): # If word does not exist if not curr.is_leaf: return False lowerCamelCase__ : str = False return len(curr.nodes ) == 0 lowerCamelCase__ : List[str] = word[index] lowerCamelCase__ : Dict = curr.nodes.get(UpperCamelCase__ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCamelCase__ : List[Any] = _delete(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCamelCase__ , 0 ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: if node.is_leaf: print(UpperCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(UpperCamelCase , word + key ) def SCREAMING_SNAKE_CASE_ () -> bool: lowerCamelCase__ : str = """banana bananas bandana band apple all beast""".split() lowerCamelCase__ : Union[str, Any] = TrieNode() root.insert_many(UpperCamelCase ) # print_words(root, "") assert all(root.find(UpperCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: print(str(UpperCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def SCREAMING_SNAKE_CASE_ () -> None: assert test_trie() def SCREAMING_SNAKE_CASE_ () -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
41
1
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : str = 0.0 for coeff in reversed(UpperCamelCase ): lowerCamelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": _A : Any =(0.0, 0.0, 5.0, 9.3, 7.0) _A : Optional[Any] =10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
41
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _A : Any ='''▁''' _A : str ={'''vocab_file''': '''spiece.model'''} _A : Any ={ '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _A : Union[str, Any] ={ '''google/pegasus-xsum''': 512, } _A : Union[str, Any] =logging.get_logger(__name__) class _lowercase ( _lowercase ): a = VOCAB_FILES_NAMES a = VOCAB_FILES_NAMES a = PRETRAINED_VOCAB_FILES_MAP a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a = ["""input_ids""", """attention_mask"""] def __init__( self: Any , UpperCamelCase__: int , UpperCamelCase__: Union[str, Any]="<pad>" , UpperCamelCase__: Optional[int]="</s>" , UpperCamelCase__: Optional[Any]="<unk>" , UpperCamelCase__: Optional[int]="<mask_2>" , UpperCamelCase__: List[str]="<mask_1>" , UpperCamelCase__: Optional[Any]=None , UpperCamelCase__: Any=103 , UpperCamelCase__: Optional[Dict[str, Any]] = None , **UpperCamelCase__: Optional[int] , ): lowerCamelCase__ : Dict = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError( F'''additional_special_tokens should be of type {type(UpperCamelCase__ )}, but is''' F''' {type(UpperCamelCase__ )}''' ) lowerCamelCase__ : Union[str, Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'''<unk_{i}>''' for i in range(len(UpperCamelCase__ ) , self.offset - 1 ) ] if len(set(UpperCamelCase__ ) ) != len(UpperCamelCase__ ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" F''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) lowerCamelCase__ : Tuple = additional_special_tokens_extended else: lowerCamelCase__ : List[Any] = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'''<unk_{i}>''' for i in range(2 , self.offset )] lowerCamelCase__ : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token_sent=UpperCamelCase__ , offset=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) lowerCamelCase__ : Optional[int] = mask_token_sent lowerCamelCase__ : Optional[Any] = vocab_file lowerCamelCase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) # add special tokens to encoder dict lowerCamelCase__ : Dict[int, str] = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowerCamelCase__ : Dict[str, int] = {v: k for k, v in self.encoder.items()} @property def lowerCamelCase_ ( self: Tuple ): return len(self.sp_model ) + self.offset def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self: List[Any] ): lowerCamelCase__ : int = self.__dict__.copy() lowerCamelCase__ : Tuple = None return state def __setstate__( self: Optional[Any] , UpperCamelCase__: List[Any] ): lowerCamelCase__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase__ : Dict = {} lowerCamelCase__ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: str ): return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: str ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowerCamelCase__ : Tuple = self.sp_model.piece_to_id(UpperCamelCase__ ) return sp_id + self.offset def lowerCamelCase_ ( self: int , UpperCamelCase__: int ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowerCamelCase__ : List[str] = self.sp_model.IdToPiece(index - self.offset ) return token def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[Any] ): lowerCamelCase__ : Any = [] lowerCamelCase__ : str = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase__ ) + token lowerCamelCase__ : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase__ ) out_string += self.sp_model.decode(UpperCamelCase__ ) return out_string.strip() def lowerCamelCase_ ( self: int , UpperCamelCase__: Optional[int]=False ): return 1 def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def lowerCamelCase_ ( self: str , UpperCamelCase__: List , UpperCamelCase__: Optional[List] = None , UpperCamelCase__: bool = False ): if already_has_special_tokens: return self._special_token_mask(UpperCamelCase__ ) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def lowerCamelCase_ ( self: Any , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str , UpperCamelCase__: Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCamelCase__ : str = os.path.join( UpperCamelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , """wb""" ) as fi: lowerCamelCase__ : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,)
41
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _A : List[Any] =logging.get_logger(__name__) _A : Dict =['''model.decoder.embed_positions.weights'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: if "emb" in name: lowerCamelCase__ : Dict = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: lowerCamelCase__ : List[str] = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: lowerCamelCase__ : List[str] = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: lowerCamelCase__ : Optional[int] = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: lowerCamelCase__ : Dict = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: lowerCamelCase__ : Optional[Any] = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: lowerCamelCase__ : Optional[Any] = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: lowerCamelCase__ : int = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple[Dict, Dict]: lowerCamelCase__ : int = list(state_dict.keys() ) lowerCamelCase__ : Tuple = {} for key in keys: lowerCamelCase__ : Any = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj lowerCamelCase__ : Union[str, Any] = val[:hidden_size, :] lowerCamelCase__ : Any = val[hidden_size : 2 * hidden_size, :] lowerCamelCase__ : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: lowerCamelCase__ : str = val else: lowerCamelCase__ : Union[str, Any] = val return state_dict, enc_dec_proj_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values lowerCamelCase__ : int = 1024 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 16 elif checkpoint == "medium": lowerCamelCase__ : Any = 1536 lowerCamelCase__ : Union[str, Any] = 48 lowerCamelCase__ : Optional[int] = 24 elif checkpoint == "large": lowerCamelCase__ : Optional[Any] = 2048 lowerCamelCase__ : Dict = 48 lowerCamelCase__ : List[Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) lowerCamelCase__ : Any = MusicgenDecoderConfig( hidden_size=UpperCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase , num_attention_heads=UpperCamelCase , ) return config @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase="cpu" ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = MusicGen.get_pretrained(UpperCamelCase , device=UpperCamelCase ) lowerCamelCase__ : List[Any] = decoder_config_from_checkpoint(UpperCamelCase ) lowerCamelCase__ : Any = fairseq_model.lm.state_dict() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = rename_state_dict( UpperCamelCase , hidden_size=decoder_config.hidden_size ) lowerCamelCase__ : str = TaEncoderModel.from_pretrained("""t5-base""" ) lowerCamelCase__ : Tuple = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) lowerCamelCase__ : Optional[int] = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection lowerCamelCase__ , lowerCamelCase__ : List[str] = decoder.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model lowerCamelCase__ : Optional[Any] = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase , audio_encoder=UpperCamelCase , decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass lowerCamelCase__ : Dict = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) lowerCamelCase__ : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor lowerCamelCase__ : str = AutoTokenizer.from_pretrained("""t5-base""" ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) lowerCamelCase__ : Optional[int] = MusicgenProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids lowerCamelCase__ : Union[str, Any] = 2048 lowerCamelCase__ : List[str] = 2048 # set other default generation config params lowerCamelCase__ : Optional[Any] = int(30 * audio_encoder.config.frame_rate ) lowerCamelCase__ : Union[str, Any] = True lowerCamelCase__ : List[Any] = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _A : List[str] =parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
41
1
'''simple docstring''' import operator as op _A : Optional[Any] ='''scaler.pt''' _A : Optional[Any] ='''pytorch_model''' _A : int ='''random_states''' _A : List[Any] ='''optimizer''' _A : Dict ='''scheduler''' _A : Dict ='''pytorch_model.bin''' _A : Optional[Any] ='''pytorch_model.bin.index.json''' _A : List[str] ='''model.safetensors''' _A : List[Any] ='''model.safetensors.index.json''' _A : str ='''1.10.2''' _A : List[Any] ='''py38''' _A : int ='''4.17.0''' _A : List[str] =['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] _A : Tuple =['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] _A : Tuple =['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] _A : Optional[int] =['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] _A : Optional[int] =['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] _A : List[Any] ='''2.0.1''' _A : str =['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] _A : List[str] =['''default''', '''reduce-overhead''', '''max-autotune'''] _A : str ={'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 _A : Tuple =[ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] _A : Any =['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] _A : Union[str, Any] =['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
41
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' import unittest import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = None , ) -> np.ndarray: lowerCamelCase__ : Tuple = np.shape(UpperCamelCase ) lowerCamelCase__ : Optional[int] = np.shape(UpperCamelCase ) lowerCamelCase__ : List[Any] = np.shape(UpperCamelCase ) if shape_a[0] != shape_b[0]: lowerCamelCase__ : List[str] = ( """Expected the same number of rows for A and B. """ f'''Instead found A of size {shape_a} and B of size {shape_b}''' ) raise ValueError(UpperCamelCase ) if shape_b[1] != shape_c[1]: lowerCamelCase__ : Tuple = ( """Expected the same number of columns for B and C. """ f'''Instead found B of size {shape_b} and C of size {shape_c}''' ) raise ValueError(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = pseudo_inv if a_inv is None: try: lowerCamelCase__ : str = np.linalg.inv(UpperCamelCase ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCamelCase__ : List[Any] = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCamelCase__ : Optional[Any] = np.array([[2, 1], [6, 3]] ) lowerCamelCase__ : Dict = schur_complement(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Any = np.block([[a, b], [b.T, c]] ) lowerCamelCase__ : List[Any] = np.linalg.det(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = np.linalg.det(UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = np.linalg.det(UpperCamelCase__ ) self.assertAlmostEqual(UpperCamelCase__ , det_a * det_s ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : List[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCamelCase__ : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCamelCase__ : str = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCamelCase__ ): schur_complement(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Dict = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowerCamelCase__ : int = np.array([[0, 3], [3, 0], [2, 3]] ) lowerCamelCase__ : List[str] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCamelCase__ ): schur_complement(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: assert column_title.isupper() lowerCamelCase__ : Dict = 0 lowerCamelCase__ : str = len(UpperCamelCase ) - 1 lowerCamelCase__ : str = 0 while index >= 0: lowerCamelCase__ : List[str] = (ord(column_title[index] ) - 64) * pow(26 , UpperCamelCase ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
41
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=7 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[str]=18 , UpperCamelCase__: Union[str, Any]=30 , UpperCamelCase__: List[str]=400 , UpperCamelCase__: Any=True , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[Any]=False , UpperCamelCase__: Tuple=True , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , ): lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Optional[int] = min_resolution lowerCamelCase__ : Optional[Any] = max_resolution lowerCamelCase__ : Union[str, Any] = do_resize lowerCamelCase__ : Optional[Any] = size if size is not None else {"""height""": 18, """width""": 20} lowerCamelCase__ : Dict = do_thumbnail lowerCamelCase__ : Optional[int] = do_align_axis lowerCamelCase__ : Any = do_pad lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Union[str, Any] = image_mean lowerCamelCase__ : Union[str, Any] = image_std def lowerCamelCase_ ( self: str ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DonutImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = DonutImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_thumbnail""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_align_long_axis""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 20} ) lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order lowerCamelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"""height""": 84, """width""": 42} ) def lowerCamelCase_ ( self: List[str] ): pass @is_flaky() def lowerCamelCase_ ( self: Union[str, Any] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Optional[int] ): # Initialize image_processing lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Dict ): # Initialize image_processing lowerCamelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
41
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> list[list[int]]: lowerCamelCase__ : list[list[int]] = [] lowerCamelCase__ : list[int] = [] lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : str = sum(UpperCamelCase ) create_state_space_tree(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) return result def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> None: if sum(UpperCamelCase ) > max_sum or (remaining_nums_sum + sum(UpperCamelCase )) < max_sum: return if sum(UpperCamelCase ) == max_sum: result.append(UpperCamelCase ) return for index in range(UpperCamelCase , len(UpperCamelCase ) ): create_state_space_tree( UpperCamelCase , UpperCamelCase , index + 1 , [*path, nums[index]] , UpperCamelCase , remaining_nums_sum - nums[index] , ) _A : Tuple =[3, 34, 4, 12, 5, 2] _A : Any =9 _A : List[str] =generate_sum_of_subsets_soln(nums, max_sum) print(*result)
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' from scipy.stats import pearsonr import datasets _A : Optional[int] =''' Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. ''' _A : Optional[int] =''' Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results[\'pearsonr\'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) [\'p-value\', \'pearsonr\'] >>> print(round(results[\'pearsonr\'], 2)) -0.74 >>> print(round(results[\'p-value\'], 2)) 0.15 ''' _A : List[str] =''' @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def lowerCamelCase_ ( self: str ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Any=False ): if return_pvalue: lowerCamelCase__ : str = pearsonr(UpperCamelCase__ , UpperCamelCase__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(UpperCamelCase__ , UpperCamelCase__ )[0] )}
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _A : Dict =0 _A : Optional[Any] =[ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _A : int =[[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _A : Optional[Any] =tuple[int, int] class _lowercase : def __init__( self: List[str] , UpperCamelCase__: int , UpperCamelCase__: int , UpperCamelCase__: int , UpperCamelCase__: int , UpperCamelCase__: int , UpperCamelCase__: Node | None , ): lowerCamelCase__ : Optional[int] = pos_x lowerCamelCase__ : List[Any] = pos_y lowerCamelCase__ : Dict = (pos_y, pos_x) lowerCamelCase__ : str = goal_x lowerCamelCase__ : Optional[int] = goal_y lowerCamelCase__ : List[str] = g_cost lowerCamelCase__ : Tuple = parent lowerCamelCase__ : Any = self.calculate_heuristic() lowerCamelCase__ : Tuple = self.g_cost + self.h_cost def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.pos_x - self.goal_x lowerCamelCase__ : Tuple = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(UpperCamelCase__ ) + abs(UpperCamelCase__ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self: List[str] , UpperCamelCase__: Node ): return self.f_cost < other.f_cost class _lowercase : def __init__( self: Tuple , UpperCamelCase__: TPosition , UpperCamelCase__: TPosition ): lowerCamelCase__ : Any = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , UpperCamelCase__ ) lowerCamelCase__ : int = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = [self.start] lowerCamelCase__ : list[Node] = [] lowerCamelCase__ : Dict = False def lowerCamelCase_ ( self: int ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() lowerCamelCase__ : int = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(UpperCamelCase__ ) self.closed_nodes.append(UpperCamelCase__ ) lowerCamelCase__ : str = self.get_successors(UpperCamelCase__ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(UpperCamelCase__ ) else: # retrieve the best current path lowerCamelCase__ : Optional[Any] = self.open_nodes.pop(self.open_nodes.index(UpperCamelCase__ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(UpperCamelCase__ ) else: self.open_nodes.append(UpperCamelCase__ ) return [self.start.pos] def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Node ): lowerCamelCase__ : Tuple = [] for action in delta: lowerCamelCase__ : Any = parent.pos_x + action[1] lowerCamelCase__ : List[str] = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(UpperCamelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( UpperCamelCase__ , UpperCamelCase__ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , UpperCamelCase__ , ) ) return successors def lowerCamelCase_ ( self: Any , UpperCamelCase__: Node | None ): lowerCamelCase__ : Optional[Any] = node lowerCamelCase__ : Optional[Any] = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) lowerCamelCase__ : List[Any] = current_node.parent path.reverse() return path class _lowercase : def __init__( self: str , UpperCamelCase__: TPosition , UpperCamelCase__: TPosition ): lowerCamelCase__ : str = AStar(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Dict = AStar(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = False def lowerCamelCase_ ( self: List[Any] ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() lowerCamelCase__ : Any = self.fwd_astar.open_nodes.pop(0 ) lowerCamelCase__ : int = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( UpperCamelCase__ , UpperCamelCase__ ) self.fwd_astar.closed_nodes.append(UpperCamelCase__ ) self.bwd_astar.closed_nodes.append(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = current_bwd_node lowerCamelCase__ : Optional[int] = current_fwd_node lowerCamelCase__ : int = { self.fwd_astar: self.fwd_astar.get_successors(UpperCamelCase__ ), self.bwd_astar: self.bwd_astar.get_successors(UpperCamelCase__ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(UpperCamelCase__ ) else: # retrieve the best current path lowerCamelCase__ : Optional[Any] = astar.open_nodes.pop( astar.open_nodes.index(UpperCamelCase__ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(UpperCamelCase__ ) else: astar.open_nodes.append(UpperCamelCase__ ) return [self.fwd_astar.start.pos] def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Node , UpperCamelCase__: Node ): lowerCamelCase__ : Any = self.fwd_astar.retrace_path(UpperCamelCase__ ) lowerCamelCase__ : Any = self.bwd_astar.retrace_path(UpperCamelCase__ ) bwd_path.pop() bwd_path.reverse() lowerCamelCase__ : int = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _A : int =(0, 0) _A : Any =(len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _A : List[Any] =time.time() _A : Optional[Any] =AStar(init, goal) _A : List[Any] =a_star.search() _A : Dict =time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') _A : List[Any] =time.time() _A : int =BidirectionalAStar(init, goal) _A : List[str] =time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
41
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowerCamelCase__ : List[Any] = torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = pipe( image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowerCamelCase__ : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase__ : Tuple = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ , lowerCamelCase__ : str = [], [] while len(UpperCamelCase ) > 1: lowerCamelCase__ , lowerCamelCase__ : int = min(UpperCamelCase ), max(UpperCamelCase ) start.append(UpperCamelCase ) end.append(UpperCamelCase ) collection.remove(UpperCamelCase ) collection.remove(UpperCamelCase ) end.reverse() return start + collection + end if __name__ == "__main__": _A : Any =input('''Enter numbers separated by a comma:\n''').strip() _A : str =[int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
41
'''simple docstring''' # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _A : Dict ='''tiny-wmt19-en-ru''' # Build # borrowed from a test _A : List[str] =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _A : str =dict(zip(vocab, range(len(vocab)))) _A : List[str] =['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: _A : Union[str, Any] =Path(tmpdirname) _A : str =build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] _A : int =build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] _A : List[Any] =build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) _A : int =FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _A : List[str] =FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _A : Union[str, Any] =FSMTForConditionalGeneration(config) print(F'num of params {tiny_model.num_parameters()}') # Test _A : List[str] =tokenizer(['''Making tiny model'''], return_tensors='''pt''') _A : Tuple =tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'Generated {mname_tiny}') # Upload # transformers-cli upload tiny-wmt19-en-ru
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : int ={ '''configuration_mobilebert''': [ '''MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileBertConfig''', '''MobileBertOnnxConfig''', ], '''tokenization_mobilebert''': ['''MobileBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : str =['''MobileBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileBertForMaskedLM''', '''MobileBertForMultipleChoice''', '''MobileBertForNextSentencePrediction''', '''MobileBertForPreTraining''', '''MobileBertForQuestionAnswering''', '''MobileBertForSequenceClassification''', '''MobileBertForTokenClassification''', '''MobileBertLayer''', '''MobileBertModel''', '''MobileBertPreTrainedModel''', '''load_tf_weights_in_mobilebert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : str =[ '''TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileBertForMaskedLM''', '''TFMobileBertForMultipleChoice''', '''TFMobileBertForNextSentencePrediction''', '''TFMobileBertForPreTraining''', '''TFMobileBertForQuestionAnswering''', '''TFMobileBertForSequenceClassification''', '''TFMobileBertForTokenClassification''', '''TFMobileBertMainLayer''', '''TFMobileBertModel''', '''TFMobileBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' import os import unittest from transformers.models.phobert.tokenization_phobert import VOCAB_FILES_NAMES, PhobertTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _lowercase ( _lowercase , unittest.TestCase ): a = PhobertTokenizer a = False def lowerCamelCase_ ( self: Optional[Any] ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCamelCase__ : List[Any] = ["""T@@""", """i""", """I""", """R@@""", """r""", """e@@"""] lowerCamelCase__ : str = dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) lowerCamelCase__ : Optional[int] = ["""#version: 0.2""", """l à</w>"""] lowerCamelCase__ : Tuple = {"""unk_token""": """<unk>"""} lowerCamelCase__ : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCamelCase__ : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'''{token} {vocab_tokens[token]}\n''' ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(UpperCamelCase__ ) ) def lowerCamelCase_ ( self: Optional[int] , **UpperCamelCase__: int ): kwargs.update(self.special_tokens_map ) return PhobertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: int ): lowerCamelCase__ : int = """Tôi là VinAI Research""" lowerCamelCase__ : Tuple = """T<unk> i <unk> <unk> <unk> <unk> <unk> <unk> I Re<unk> e<unk> <unk> <unk> <unk>""" return input_text, output_text def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Dict = PhobertTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCamelCase__ : Optional[Any] = """Tôi là VinAI Research""" lowerCamelCase__ : List[str] = """T@@ ô@@ i l@@ à V@@ i@@ n@@ A@@ I R@@ e@@ s@@ e@@ a@@ r@@ c@@ h""".split() lowerCamelCase__ : str = tokenizer.tokenize(UpperCamelCase__ ) print(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Any = tokens + [tokenizer.unk_token] lowerCamelCase__ : Any = [4, 3, 5, 3, 3, 3, 3, 3, 3, 6, 7, 9, 3, 9, 3, 3, 3, 3, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , UpperCamelCase__ )
41
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: Optional[Any]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: str=0.6 , UpperCamelCase__: str=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Optional[Any] = batch_size lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Optional[Any] = is_training lowerCamelCase__ : int = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : List[str] = mask_ratio lowerCamelCase__ : Optional[int] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Any = (image_size // patch_size) ** 2 lowerCamelCase__ : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches lowerCamelCase__ : List[str] = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Union[str, Any] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Optional[int] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = TFViTMAEModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : int = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : str = outputs_dict[0].numpy() lowerCamelCase__ : Optional[int] = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def lowerCamelCase_ ( self: Dict ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__: int ): lowerCamelCase__ : Optional[int] = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): lowerCamelCase__ : List[str] = v.numpy() else: lowerCamelCase__ : Union[str, Any] = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = prepare_numpy_arrays(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : Any = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: str ): # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase__ : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : Tuple = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[Any] = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } lowerCamelCase__ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase__ : List[str] = main_layer_class(UpperCamelCase__ ) lowerCamelCase__ : int = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase__ : List[str] = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) lowerCamelCase__ : Union[str, Any] = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : int = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) lowerCamelCase__ : int = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : Any = outputs.last_hidden_state.numpy() lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : int = outputs.logits.numpy() lowerCamelCase__ : Dict = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) lowerCamelCase__ : Dict = model_class.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : str = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase__ : Optional[Any] = 0 else: lowerCamelCase__ : Union[str, Any] = after_outputs["""logits"""].numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) def lowerCamelCase_ ( self: Any ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Optional[int] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) lowerCamelCase__ : int = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase__ : int = model_class.from_config(model.config ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @slow def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[Any] ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: List[str] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Tuple = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase__ : str = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Any = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1e-4 )
41
1
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins _A : Union[str, Any] =['''tests.fixtures.files''', '''tests.fixtures.hub''', '''tests.fixtures.fsspec'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[Any]: # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ["""integration""", """unit"""] ): continue item.add_marker(pytest.mark.unit ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: config.addinivalue_line("""markers""" , """torchaudio_latest: mark test to run with torchaudio>=0.12""" ) @pytest.fixture(autouse=UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Any: # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? lowerCamelCase__ : Dict = tmp_path_factory.getbasetemp() / """cache""" lowerCamelCase__ : Optional[Any] = test_hf_cache_home / """datasets""" lowerCamelCase__ : List[Any] = test_hf_cache_home / """metrics""" lowerCamelCase__ : List[str] = test_hf_cache_home / """modules""" monkeypatch.setattr("""datasets.config.HF_DATASETS_CACHE""" , str(UpperCamelCase ) ) monkeypatch.setattr("""datasets.config.HF_METRICS_CACHE""" , str(UpperCamelCase ) ) monkeypatch.setattr("""datasets.config.HF_MODULES_CACHE""" , str(UpperCamelCase ) ) lowerCamelCase__ : str = test_hf_datasets_cache / """downloads""" monkeypatch.setattr("""datasets.config.DOWNLOADED_DATASETS_PATH""" , str(UpperCamelCase ) ) lowerCamelCase__ : Optional[Any] = test_hf_datasets_cache / """downloads""" / """extracted""" monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(UpperCamelCase ) ) @pytest.fixture(autouse=UpperCamelCase , scope="""session""" ) def SCREAMING_SNAKE_CASE_ () -> int: datasets.disable_progress_bar() @pytest.fixture(autouse=UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Union[str, Any]: # don't take tests into account when counting downloads monkeypatch.setattr("""datasets.config.HF_UPDATE_DOWNLOAD_COUNTS""" , UpperCamelCase ) @pytest.fixture def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Optional[int]: # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr("""sqlalchemy.util.deprecations.SILENCE_UBER_WARNING""" , UpperCamelCase )
41
'''simple docstring''' class _lowercase : def __init__( self: Tuple , UpperCamelCase__: list[int] ): lowerCamelCase__ : Union[str, Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [0] * len_array if len_array > 0: lowerCamelCase__ : Union[str, Any] = array[0] for i in range(1 , UpperCamelCase__ ): lowerCamelCase__ : Tuple = self.prefix_sum[i - 1] + array[i] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: int ): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : Dict = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(UpperCamelCase__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() _A : Dict =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: lowerCamelCase__ : Union[str, Any] = MobileNetVaConfig(layer_norm_eps=0.001 ) if "_quant" in model_name: raise ValueError("""Quantized models are not supported.""" ) lowerCamelCase__ : Optional[int] = re.match(r"""^mobilenet_v1_([^_]*)_([^_]*)$""" , UpperCamelCase ) if matches: lowerCamelCase__ : Optional[int] = float(matches[1] ) lowerCamelCase__ : Optional[int] = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". lowerCamelCase__ : List[Any] = 1001 lowerCamelCase__ : Any = """imagenet-1k-id2label.json""" lowerCamelCase__ : Union[str, Any] = """huggingface/label-files""" lowerCamelCase__ : List[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : str = {int(UpperCamelCase ) + 1: v for k, v in idalabel.items()} lowerCamelCase__ : Dict = """background""" lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : Dict = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : List[str] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase=False ) -> Tuple: lowerCamelCase__ : Union[str, Any] = get_mobilenet_va_config(UpperCamelCase ) # Load 🤗 model lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor lowerCamelCase__ : Optional[Any] = MobileNetVaImageProcessor( crop_size={"""width""": config.image_size, """height""": config.image_size} , size={"""shortest_edge""": config.image_size + 32} , ) lowerCamelCase__ : Any = image_processor(images=prepare_img() , return_tensors="""pt""" ) lowerCamelCase__ : str = model(**UpperCamelCase ) lowerCamelCase__ : Dict = outputs.logits assert logits.shape == (1, 1001) if model_name == "mobilenet_v1_1.0_224": lowerCamelCase__ : Optional[Any] = torch.tensor([-4.1739, -1.1233, 3.1205] ) elif model_name == "mobilenet_v1_0.75_192": lowerCamelCase__ : Union[str, Any] = torch.tensor([-3.9440, -2.3141, -0.3333] ) else: lowerCamelCase__ : Tuple = None if expected_logits is not None: assert torch.allclose(logits[0, :3] , UpperCamelCase , atol=1E-4 ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase ) if push_to_hub: print("""Pushing to the hub...""" ) lowerCamelCase__ : Optional[Any] = """google/""" + model_name image_processor.push_to_hub(UpperCamelCase ) model.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Any =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''mobilenet_v1_1.0_224''', type=str, help='''Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.''', ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original TensorFlow checkpoint (.ckpt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _A : Optional[Any] =parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A : Dict ={ '''configuration_swinv2''': ['''SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Swinv2Config'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Swinv2ForImageClassification''', '''Swinv2ForMaskedImageModeling''', '''Swinv2Model''', '''Swinv2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Union[str, Any]: if isinstance(UpperCamelCase , collections.abc.Iterable ): return x return (x, x) @require_flax class _lowercase : def lowerCamelCase_ ( self: Dict , UpperCamelCase__: str , UpperCamelCase__: Optional[Any] ): pass def lowerCamelCase_ ( self: Dict ): pass def lowerCamelCase_ ( self: Tuple ): pass def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: np.ndarray , UpperCamelCase__: np.ndarray , UpperCamelCase__: float ): lowerCamelCase__ : List[str] = np.abs((a - b) ).max() self.assertLessEqual(UpperCamelCase__ , UpperCamelCase__ , F'''Difference between torch and flax is {diff} (>= {tol}).''' ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[Any] , UpperCamelCase__: List[str] , UpperCamelCase__: Tuple , UpperCamelCase__: int=None , **UpperCamelCase__: Any ): lowerCamelCase__ : Dict = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Tuple = FlaxVisionTextDualEncoderModel(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], config.projection_dim) ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str]=None , **UpperCamelCase__: Optional[int] ): lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.get_vision_text_model(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Dict = {"""vision_model""": vision_model, """text_model""": text_model} lowerCamelCase__ : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = model(input_ids=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any]=None , **UpperCamelCase__: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Tuple = self.get_vision_text_model(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = {"""vision_model""": vision_model, """text_model""": text_model} lowerCamelCase__ : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase__ ) lowerCamelCase__ : Dict = model(input_ids=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) lowerCamelCase__ : Tuple = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(input_ids=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = after_output[0] lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-3 ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any , UpperCamelCase__: Optional[Any]=None , **UpperCamelCase__: Any ): lowerCamelCase__ , lowerCamelCase__ : str = self.get_vision_text_model(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = {"""vision_model""": vision_model, """text_model""": text_model} lowerCamelCase__ : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**UpperCamelCase__ ) lowerCamelCase__ : str = model( input_ids=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , output_attentions=UpperCamelCase__ ) lowerCamelCase__ : int = output.vision_model_output.attentions self.assertEqual(len(UpperCamelCase__ ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase__ : Dict = to_atuple(vision_model.config.image_size ) lowerCamelCase__ : Union[str, Any] = to_atuple(vision_model.config.patch_size ) lowerCamelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowerCamelCase__ : Union[str, Any] = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) lowerCamelCase__ : List[str] = output.text_model_output.attentions self.assertEqual(len(UpperCamelCase__ ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): pt_model.to(UpperCamelCase__ ) pt_model.eval() # prepare inputs lowerCamelCase__ : List[str] = inputs_dict lowerCamelCase__ : Optional[Any] = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): lowerCamelCase__ : List[str] = pt_model(**UpperCamelCase__ ).to_tuple() lowerCamelCase__ : List[Any] = fx_model(**UpperCamelCase__ ).to_tuple() self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(UpperCamelCase__ , pt_output.numpy() , 4e-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Dict = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase__ , from_pt=UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = fx_model_loaded(**UpperCamelCase__ ).to_tuple() self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(UpperCamelCase__ , pt_output.numpy() , 4e-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = VisionTextDualEncoderModel.from_pretrained(UpperCamelCase__ , from_flax=UpperCamelCase__ ) pt_model_loaded.to(UpperCamelCase__ ) pt_model_loaded.eval() with torch.no_grad(): lowerCamelCase__ : Optional[Any] = pt_model_loaded(**UpperCamelCase__ ).to_tuple() self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(UpperCamelCase__ , pt_output_loaded.numpy() , 4e-2 ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Any , UpperCamelCase__: int , UpperCamelCase__: Any ): lowerCamelCase__ : int = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = VisionTextDualEncoderModel(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = FlaxVisionTextDualEncoderModel(UpperCamelCase__ ) lowerCamelCase__ : str = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , UpperCamelCase__ ) lowerCamelCase__ : Tuple = fx_state self.check_pt_flax_equivalence(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: Dict , UpperCamelCase__: int , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : int = VisionTextDualEncoderModel(UpperCamelCase__ ) lowerCamelCase__ : Tuple = FlaxVisionTextDualEncoderModel(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = load_flax_weights_in_pytorch_model(UpperCamelCase__ , fx_model.params ) self.check_pt_flax_equivalence(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Tuple = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**UpperCamelCase__ ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() self.check_save_load(**UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[str] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**UpperCamelCase__ ) @is_pt_flax_cross_test def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Any = self.prepare_config_and_inputs() lowerCamelCase__ : int = config_inputs_dict.pop("""vision_config""" ) lowerCamelCase__ : str = config_inputs_dict.pop("""text_config""" ) lowerCamelCase__ : Dict = config_inputs_dict self.check_equivalence_pt_to_flax(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) self.check_equivalence_flax_to_pt(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ , lowerCamelCase__ : Any = self.get_pretrained_model_and_inputs() lowerCamelCase__ : Dict = model_a(**UpperCamelCase__ ) lowerCamelCase__ : str = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Any = model_a(**UpperCamelCase__ ) lowerCamelCase__ : Any = after_outputs[0] lowerCamelCase__ : List[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) @require_flax class _lowercase ( _lowercase , unittest.TestCase ): def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( """hf-internal-testing/tiny-random-vit""" , """hf-internal-testing/tiny-bert""" , vision_from_pt=UpperCamelCase__ , text_from_pt=UpperCamelCase__ , ) lowerCamelCase__ : Optional[int] = 13 lowerCamelCase__ : Optional[Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCamelCase__ : List[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) lowerCamelCase__ : Any = random_attention_mask([batch_size, 4] ) lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Tuple , UpperCamelCase__: Tuple ): lowerCamelCase__ : Dict = FlaxViTModel(UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = FlaxBertModel(UpperCamelCase__ ) return vision_model, text_model def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : List[Any] = FlaxViTModelTester(self ) lowerCamelCase__ : List[Any] = FlaxBertModelTester(self ) lowerCamelCase__ : List[Any] = vit_model_tester.prepare_config_and_inputs() lowerCamelCase__ : str = bert_model_tester.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = vision_config_and_inputs lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class _lowercase ( _lowercase , unittest.TestCase ): def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( """hf-internal-testing/tiny-random-clip""" , """hf-internal-testing/tiny-bert""" , vision_from_pt=UpperCamelCase__ , text_from_pt=UpperCamelCase__ , ) lowerCamelCase__ : str = 13 lowerCamelCase__ : Tuple = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) lowerCamelCase__ : Optional[Any] = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) lowerCamelCase__ : str = random_attention_mask([batch_size, 4] ) lowerCamelCase__ : int = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: List[Any] ): lowerCamelCase__ : List[Any] = FlaxCLIPVisionModel(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = FlaxBertModel(UpperCamelCase__ ) return vision_model, text_model def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = FlaxCLIPVisionModelTester(self ) lowerCamelCase__ : Tuple = FlaxBertModelTester(self ) lowerCamelCase__ : Any = clip_model_tester.prepare_config_and_inputs() lowerCamelCase__ : List[str] = bert_model_tester.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ : str = vision_config_and_inputs lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class _lowercase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Optional[Any] = FlaxVisionTextDualEncoderModel.from_pretrained("""clip-italian/clip-italian""" , logit_scale_init_value=1.0 ) lowerCamelCase__ : str = VisionTextDualEncoderProcessor.from_pretrained("""clip-italian/clip-italian""" ) lowerCamelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) lowerCamelCase__ : Dict = processor( text=["""una foto di un gatto""", """una foto di un cane"""] , images=UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase__ ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) lowerCamelCase__ : str = np.array([[1.2_284_727, 0.3_104_122]] ) self.assertTrue(np.allclose(outputs.logits_per_image , UpperCamelCase__ , atol=1e-3 ) )
41
'''simple docstring''' from __future__ import annotations _A : Any ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class _lowercase : def __init__( self: Tuple , UpperCamelCase__: dict[str, list[str]] , UpperCamelCase__: str ): lowerCamelCase__ : str = graph # mapping node to its parent in resulting breadth first tree lowerCamelCase__ : dict[str, str | None] = {} lowerCamelCase__ : Any = source_vertex def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = {self.source_vertex} lowerCamelCase__ : List[str] = None lowerCamelCase__ : Tuple = [self.source_vertex] # first in first out queue while queue: lowerCamelCase__ : Tuple = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase__ ) lowerCamelCase__ : List[str] = vertex queue.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): if target_vertex == self.source_vertex: return self.source_vertex lowerCamelCase__ : Tuple = self.parent.get(UpperCamelCase__ ) if target_vertex_parent is None: lowerCamelCase__ : int = ( F'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(UpperCamelCase__ ) return self.shortest_path(UpperCamelCase__ ) + F'''->{target_vertex}''' if __name__ == "__main__": _A : int =Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
41
1
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class _lowercase : def __init__( self: int , UpperCamelCase__: int ): lowerCamelCase__ : List[str] = str(id_ ) lowerCamelCase__ : Union[str, Any] = None lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = [] lowerCamelCase__ : int = {} # {vertex:distance} def __lt__( self: List[str] , UpperCamelCase__: Dict ): return self.key < other.key def __repr__( self: str ): return self.id def lowerCamelCase_ ( self: Any , UpperCamelCase__: int ): self.neighbors.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : Any = weight def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , UpperCamelCase ) graph[b - 1].add_edge(graph[a - 1] , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> list: lowerCamelCase__ : List[Any] = [] for u in graph: lowerCamelCase__ : Optional[int] = math.inf lowerCamelCase__ : List[str] = None lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Dict = graph[:] while q: lowerCamelCase__ : Dict = min(UpperCamelCase ) q.remove(UpperCamelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): lowerCamelCase__ : str = u lowerCamelCase__ : Dict = u.edges[v.id] for i in range(1 , len(UpperCamelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Iterator[tuple]: for u in graph: lowerCamelCase__ : Union[str, Any] = math.inf lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[Any] = list(UpperCamelCase ) hq.heapify(UpperCamelCase ) while h: lowerCamelCase__ : Dict = hq.heappop(UpperCamelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): lowerCamelCase__ : int = u lowerCamelCase__ : Optional[int] = u.edges[v.id] hq.heapify(UpperCamelCase ) for i in range(1 , len(UpperCamelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def SCREAMING_SNAKE_CASE_ () -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : str = 0.0 for coeff in reversed(UpperCamelCase ): lowerCamelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": _A : Any =(0.0, 0.0, 5.0, 9.3, 7.0) _A : Optional[Any] =10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
41
1
'''simple docstring''' # Function to print upper half of diamond (pyramid) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> List[str]: for i in range(0 , UpperCamelCase ): for _ in range(0 , n - i - 1 ): # printing spaces print(""" """ , end="""""" ) for _ in range(0 , i + 1 ): # printing stars print("""* """ , end="""""" ) print() def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: for i in range(UpperCamelCase , 0 , -1 ): for _ in range(UpperCamelCase , 0 , -1 ): # printing stars print("""* """ , end="""""" ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(""" """ , end="""""" ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: if n <= 0: print(""" ... .... nothing printing :(""" ) return floyd(UpperCamelCase ) # upper half reverse_floyd(UpperCamelCase ) # lower half if __name__ == "__main__": print(r'''| /\ | |- | |- |--| |\ /| |-''') print(r'''|/ \| |- |_ |_ |__| | \/ | |_''') _A : List[Any] =1 while K: _A : Optional[int] =int(input('''enter the number and , and see the magic : ''')) print() pretty_print(user_number) _A : Union[str, Any] =int(input('''press 0 to exit... and 1 to continue...''')) print('''Good Bye...''')
41
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _A : List[Any] ='''pt''' elif is_tf_available(): _A : Any ='''tf''' else: _A : List[str] ='''jax''' class _lowercase ( _lowercase , unittest.TestCase ): a = ByTaTokenizer a = False def lowerCamelCase_ ( self: str ): super().setUp() lowerCamelCase__ : str = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ByTaTokenizer.from_pretrained("""google/byt5-small""" ) def lowerCamelCase_ ( self: Any , **UpperCamelCase__: Tuple ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Any=False , UpperCamelCase__: Union[str, Any]=20 , UpperCamelCase__: Optional[int]=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase__ ) ): try: lowerCamelCase__ : Tuple = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Union[str, Any] = list(filter(lambda UpperCamelCase__ : re.match(R"""^[ a-zA-Z]+$""" , t[1] ) , UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = list(filter(lambda UpperCamelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase__ ) , UpperCamelCase__ ) ) if max_length is not None and len(UpperCamelCase__ ) > max_length: lowerCamelCase__ : Dict = toks[:max_length] if min_length is not None and len(UpperCamelCase__ ) < min_length and len(UpperCamelCase__ ) > 0: while len(UpperCamelCase__ ) < min_length: lowerCamelCase__ : Optional[Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : Union[str, Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : Union[str, Any] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) if " " not in output_txt and len(UpperCamelCase__ ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase__ ) ) if with_prefix_space: lowerCamelCase__ : str = """ """ + output_txt lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) return output_txt, output_ids def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : Union[str, Any] = tokenizer(["""hi</s>""", """I went to the gym</s>""", """</s>"""] ) lowerCamelCase__ : Optional[int] = tokenizer(["""hi""", """I went to the gym""", """"""] ) self.assertListEqual(batch_with_eos_added["""input_ids"""] , batch_without_eos_added["""input_ids"""] ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[Any] = self.ta_base_tokenizer lowerCamelCase__ : Dict = """Unicode €.""" lowerCamelCase__ : List[Any] = tokenizer(UpperCamelCase__ ) lowerCamelCase__ : List[str] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : Tuple = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """Unicode €.</s>""" ) lowerCamelCase__ : List[Any] = tokenizer("""e è é ê ë""" ) lowerCamelCase__ : List[Any] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : str = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """e è é ê ë</s>""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """e è é ê ë</s>""" ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : int = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off lowerCamelCase__ : List[str] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on lowerCamelCase__ : int = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) if FRAMEWORK != "jax": lowerCamelCase__ : Any = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : str = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase__ : Tuple = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase__ ) self.assertNotIn("""decoder_attention_mask""" , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = [ """Summary of the text.""", """Another summary.""", ] lowerCamelCase__ : Union[str, Any] = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.ta_base_tokenizer lowerCamelCase__ : str = ["""A long paragraph for summarization. </s>"""] lowerCamelCase__ : Optional[Any] = ["""Summary of the text. </s>"""] # fmt: off lowerCamelCase__ : Tuple = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] lowerCamelCase__ : Any = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ , text_target=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , batch["""input_ids"""][0] ) self.assertEqual(UpperCamelCase__ , batch["""labels"""][0] ) def lowerCamelCase_ ( self: Optional[int] ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : int = tempfile.mkdtemp() lowerCamelCase__ : List[str] = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Any = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) lowerCamelCase__ : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Any = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Union[str, Any] = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[Any] = json.load(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = [F'''<extra_id_{i}>''' for i in range(125 )] lowerCamelCase__ : int = added_tokens_extra_ids + [ """an_additional_special_token""" ] lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : Dict = tokenizer_class.from_pretrained( UpperCamelCase__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=UpperCamelCase__ )] lowerCamelCase__ : Any = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : str = tokenizer_class.from_pretrained(UpperCamelCase__ ) self.assertTrue(tokenizer.decode([255] ) == """""" ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: List[str] ): pass def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: int ): # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens lowerCamelCase__ : Dict = self.get_tokenizers(fast=UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Dict = ["""t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """x""", """t""", """</s>"""] lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : str = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] lowerCamelCase__ : str = 0 lowerCamelCase__ : Any = tokenizer.convert_ids_to_tokens( UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) for attr in attributes_list: setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [] ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [token_id_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [token_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [token_id_to_test_setters] )
41
1
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: lowerCamelCase__ : Dict = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """decoder.output_projection.weight""", """_float_tensor""", """encoder.embed_positions._float_tensor""", """decoder.embed_positions._float_tensor""", ] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Optional[Any]: lowerCamelCase__ : Tuple = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: lowerCamelCase__ : Tuple = s_dict.pop(UpperCamelCase ) elif "subsample" in key: lowerCamelCase__ : List[Any] = s_dict.pop(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ , lowerCamelCase__ : Tuple = emb.weight.shape lowerCamelCase__ : int = nn.Linear(UpperCamelCase , UpperCamelCase , bias=UpperCamelCase ) lowerCamelCase__ : Optional[Any] = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : int = torch.load(UpperCamelCase , map_location="""cpu""" ) lowerCamelCase__ : List[Any] = mam_aaa["""args"""] lowerCamelCase__ : Union[str, Any] = mam_aaa["""model"""] lowerCamelCase__ : List[str] = state_dict["""decoder.output_projection.weight"""] remove_ignore_keys_(UpperCamelCase ) rename_keys(UpperCamelCase ) lowerCamelCase__ : List[Any] = state_dict["""decoder.embed_tokens.weight"""].shape[0] lowerCamelCase__ : int = args.share_decoder_input_output_embed lowerCamelCase__ : Optional[int] = [int(UpperCamelCase ) for i in args.conv_kernel_sizes.split(""",""" )] lowerCamelCase__ : List[Any] = SpeechaTextConfig( vocab_size=UpperCamelCase , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="""relu""" , num_conv_layers=len(UpperCamelCase ) , conv_channels=args.conv_channels , conv_kernel_sizes=UpperCamelCase , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=UpperCamelCase , num_beams=5 , max_length=200 , use_cache=UpperCamelCase , decoder_start_token_id=2 , early_stopping=UpperCamelCase , ) lowerCamelCase__ : List[Any] = SpeechaTextForConditionalGeneration(UpperCamelCase ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = model.model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) if len(UpperCamelCase ) > 0 and not set(UpperCamelCase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( """Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,""" f''' but all the following weights are missing {missing}''' ) if tie_embeds: lowerCamelCase__ : Union[str, Any] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: lowerCamelCase__ : Optional[Any] = lm_head_weights model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : Optional[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument('''--fairseq_path''', type=str, help='''Path to the fairseq model (.pt) file.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _A : str =parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
41
'''simple docstring''' from heapq import heappop, heappush import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> tuple[float | int, list[tuple[int, int]]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = grid.shape lowerCamelCase__ : List[str] = [-1, 1, 0, 0] lowerCamelCase__ : Dict = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] lowerCamelCase__ , lowerCamelCase__ : Any = [(0, source)], set() lowerCamelCase__ : Tuple = np.full((rows, cols) , np.inf ) lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[int] = np.empty((rows, cols) , dtype=UpperCamelCase ) lowerCamelCase__ : str = None while queue: ((lowerCamelCase__) , (lowerCamelCase__)) : List[str] = heappop(UpperCamelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: lowerCamelCase__ : Optional[int] = [] while (x, y) != source: path.append((x, y) ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = predecessors[x, y] path.append(UpperCamelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCamelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: lowerCamelCase__ : Any = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCamelCase , (dist + 1, (nx, ny)) ) lowerCamelCase__ : Union[str, Any] = dist + 1 lowerCamelCase__ : List[str] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class _lowercase ( _lowercase ): a = (DDPMParallelScheduler,) def lowerCamelCase_ ( self: Union[str, Any] , **UpperCamelCase__: str ): lowerCamelCase__ : str = { """num_train_timesteps""": 1_000, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", """variance_type""": """fixed_small""", """clip_sample""": True, } config.update(**UpperCamelCase__ ) return config def lowerCamelCase_ ( self: Tuple ): for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=UpperCamelCase__ , beta_end=UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.check_over_configs(thresholding=UpperCamelCase__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=UpperCamelCase__ , prediction_type=UpperCamelCase__ , sample_max_value=UpperCamelCase__ , ) def lowerCamelCase_ ( self: str ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): for t in [0, 500, 999]: self.check_over_forward(time_step=UpperCamelCase__ ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : int = self.scheduler_classes[0] lowerCamelCase__ : Union[str, Any] = self.get_scheduler_config() lowerCamelCase__ : Union[str, Any] = scheduler_class(**UpperCamelCase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.00_979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1e-5 def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = self.scheduler_classes[0] lowerCamelCase__ : List[Any] = self.get_scheduler_config() lowerCamelCase__ : List[str] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : str = len(UpperCamelCase__ ) lowerCamelCase__ : str = self.dummy_model() lowerCamelCase__ : int = self.dummy_sample_deter lowerCamelCase__ : Optional[int] = self.dummy_sample_deter + 0.1 lowerCamelCase__ : Optional[int] = self.dummy_sample_deter - 0.1 lowerCamelCase__ : Union[str, Any] = samplea.shape[0] lowerCamelCase__ : Union[str, Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCamelCase__ : str = torch.arange(UpperCamelCase__ )[0:3, None].repeat(1 , UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCamelCase__ : Dict = scheduler.batch_step_no_noise(UpperCamelCase__ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCamelCase__ : Union[str, Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 1_153.1_833 ) < 1e-2 assert abs(result_mean.item() - 0.5_005 ) < 1e-3 def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[Any] = self.scheduler_classes[0] lowerCamelCase__ : Dict = self.get_scheduler_config() lowerCamelCase__ : List[Any] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = self.dummy_model() lowerCamelCase__ : int = self.dummy_sample_deter lowerCamelCase__ : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(UpperCamelCase__ ) ): # 1. predict noise residual lowerCamelCase__ : Dict = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 lowerCamelCase__ : Any = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample lowerCamelCase__ : List[str] = pred_prev_sample lowerCamelCase__ : List[Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : Optional[int] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 258.9_606 ) < 1e-2 assert abs(result_mean.item() - 0.3_372 ) < 1e-3 def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCamelCase__ : Any = self.get_scheduler_config(prediction_type="""v_prediction""" ) lowerCamelCase__ : Any = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : int = len(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = self.dummy_model() lowerCamelCase__ : List[Any] = self.dummy_sample_deter lowerCamelCase__ : List[str] = torch.manual_seed(0 ) for t in reversed(range(UpperCamelCase__ ) ): # 1. predict noise residual lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 lowerCamelCase__ : List[str] = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample lowerCamelCase__ : List[Any] = pred_prev_sample lowerCamelCase__ : Union[str, Any] = torch.sum(torch.abs(UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 202.0_296 ) < 1e-2 assert abs(result_mean.item() - 0.2_631 ) < 1e-3 def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Dict = self.scheduler_classes[0] lowerCamelCase__ : List[str] = self.get_scheduler_config() lowerCamelCase__ : Optional[int] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=UpperCamelCase__ ) lowerCamelCase__ : Any = scheduler.timesteps for i, timestep in enumerate(UpperCamelCase__ ): if i == len(UpperCamelCase__ ) - 1: lowerCamelCase__ : List[str] = -1 else: lowerCamelCase__ : int = timesteps[i + 1] lowerCamelCase__ : List[Any] = scheduler.previous_timestep(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = prev_t.item() self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Optional[int] = self.scheduler_classes[0] lowerCamelCase__ : Union[str, Any] = self.get_scheduler_config() lowerCamelCase__ : Any = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = [100, 87, 50, 51, 0] with self.assertRaises(UpperCamelCase__ , msg="""`custom_timesteps` must be in descending order.""" ): scheduler.set_timesteps(timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Tuple = self.scheduler_classes[0] lowerCamelCase__ : Dict = self.get_scheduler_config() lowerCamelCase__ : str = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : Dict = [100, 87, 50, 1, 0] lowerCamelCase__ : List[str] = len(UpperCamelCase__ ) with self.assertRaises(UpperCamelCase__ , msg="""Can only pass one of `num_inference_steps` or `custom_timesteps`.""" ): scheduler.set_timesteps(num_inference_steps=UpperCamelCase__ , timesteps=UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCamelCase__ : Tuple = self.get_scheduler_config() lowerCamelCase__ : List[Any] = scheduler_class(**UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [scheduler.config.num_train_timesteps] with self.assertRaises( UpperCamelCase__ , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=UpperCamelCase__ )
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class _lowercase : def __init__( self: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any]=13 , UpperCamelCase__: Optional[Any]=2 , UpperCamelCase__: List[str]=24 , UpperCamelCase__: Optional[int]=16 , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Optional[int]=True , UpperCamelCase__: Any=32 , UpperCamelCase__: Tuple=5 , UpperCamelCase__: Union[str, Any]=4 , UpperCamelCase__: str=37 , UpperCamelCase__: Any="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: Any=0.1 , UpperCamelCase__: str=10 , UpperCamelCase__: int=0.02 , UpperCamelCase__: str=None , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Optional[Any]=2 , ): lowerCamelCase__ : Dict = parent lowerCamelCase__ : List[Any] = batch_size lowerCamelCase__ : List[str] = patch_size lowerCamelCase__ : Union[str, Any] = max_length lowerCamelCase__ : Union[str, Any] = num_mel_bins lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Union[str, Any] = use_labels lowerCamelCase__ : str = hidden_size lowerCamelCase__ : Dict = num_hidden_layers lowerCamelCase__ : Dict = num_attention_heads lowerCamelCase__ : int = intermediate_size lowerCamelCase__ : Optional[int] = hidden_act lowerCamelCase__ : List[str] = hidden_dropout_prob lowerCamelCase__ : str = attention_probs_dropout_prob lowerCamelCase__ : Union[str, Any] = type_sequence_label_size lowerCamelCase__ : Dict = initializer_range lowerCamelCase__ : Any = scope lowerCamelCase__ : Any = frequency_stride lowerCamelCase__ : Optional[Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowerCamelCase__ : Optional[Any] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowerCamelCase__ : List[Any] = (self.max_length - self.patch_size) // self.time_stride + 1 lowerCamelCase__ : Dict = frequency_out_dimension * time_out_dimension lowerCamelCase__ : str = num_patches + 2 def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[Any] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowerCamelCase__ : Any = None if self.use_labels: lowerCamelCase__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : List[str] = self.get_config() return config, input_values, labels def lowerCamelCase_ ( self: Union[str, Any] ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Tuple ): lowerCamelCase__ : int = ASTModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Tuple = config_and_inputs lowerCamelCase__ : Tuple = {"""input_values""": input_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) a = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[int] ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCamelCase_ ( self: str ): lowerCamelCase__ : str = ASTModelTester(self ) lowerCamelCase__ : List[str] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Any ): self.config_tester.run_common_tests() @unittest.skip(reason="""AST does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Dict ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Any = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , nn.Linear ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Tuple = [*signature.parameters.keys()] lowerCamelCase__ : List[Any] = ["""input_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Union[str, Any] = ASTModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[Any] = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = torchaudio.load(UpperCamelCase ) return audio, sampling_rate @require_torch @require_torchaudio class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: str ): return ( ASTFeatureExtractor.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ) if is_torchaudio_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.default_feature_extractor lowerCamelCase__ : List[Any] = ASTForAudioClassification.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self.default_feature_extractor lowerCamelCase__ , lowerCamelCase__ : str = prepare_audio() lowerCamelCase__ : Optional[int] = audio.squeeze().numpy() lowerCamelCase__ : Optional[int] = feature_extractor(UpperCamelCase__ , sampling_rate=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Tuple = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
1
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _A : int =logging.get_logger(__name__) _A : int ={ '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/config.json''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/config.json''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/config.json''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/config.json''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/config.json''', } class _lowercase ( _lowercase ): a = """t5""" a = ["""past_key_values"""] a = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self: List[str] , UpperCamelCase__: str=32_128 , UpperCamelCase__: Union[str, Any]=512 , UpperCamelCase__: str=64 , UpperCamelCase__: List[Any]=2_048 , UpperCamelCase__: Union[str, Any]=6 , UpperCamelCase__: List[str]=None , UpperCamelCase__: str=8 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Optional[int]=128 , UpperCamelCase__: Any=0.1 , UpperCamelCase__: List[str]=1e-6 , UpperCamelCase__: Dict=1.0 , UpperCamelCase__: Optional[int]="relu" , UpperCamelCase__: List[str]=True , UpperCamelCase__: List[str]=True , UpperCamelCase__: List[str]=0 , UpperCamelCase__: Union[str, Any]=1 , **UpperCamelCase__: Tuple , ): lowerCamelCase__ : Dict = vocab_size lowerCamelCase__ : List[str] = d_model lowerCamelCase__ : int = d_kv lowerCamelCase__ : Dict = d_ff lowerCamelCase__ : Any = num_layers lowerCamelCase__ : str = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCamelCase__ : Tuple = num_heads lowerCamelCase__ : Dict = relative_attention_num_buckets lowerCamelCase__ : Any = relative_attention_max_distance lowerCamelCase__ : Union[str, Any] = dropout_rate lowerCamelCase__ : Optional[int] = layer_norm_epsilon lowerCamelCase__ : str = initializer_factor lowerCamelCase__ : Dict = feed_forward_proj lowerCamelCase__ : Optional[Any] = use_cache lowerCamelCase__ : List[str] = self.feed_forward_proj.split("""-""" ) lowerCamelCase__ : Optional[int] = act_info[-1] lowerCamelCase__ : Any = act_info[0] == """gated""" if len(UpperCamelCase__ ) > 1 and act_info[0] != "gated" or len(UpperCamelCase__ ) > 2: raise ValueError( F'''`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.''' """Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. """ """'gated-gelu' or 'relu'""" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCamelCase__ : str = """gelu_new""" super().__init__( pad_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , is_encoder_decoder=UpperCamelCase__ , **UpperCamelCase__ , ) class _lowercase ( _lowercase ): @property def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = { """input_ids""": {0: """batch""", 1: """encoder_sequence"""}, """attention_mask""": {0: """batch""", 1: """encoder_sequence"""}, } if self.use_past: lowerCamelCase__ : str = """past_encoder_sequence + sequence""" lowerCamelCase__ : Union[str, Any] = {0: """batch"""} lowerCamelCase__ : Tuple = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: lowerCamelCase__ : Union[str, Any] = {0: """batch""", 1: """decoder_sequence"""} lowerCamelCase__ : Tuple = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(UpperCamelCase__ , direction="""inputs""" ) return common_inputs @property def lowerCamelCase_ ( self: str ): return 13
41
'''simple docstring''' class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : dict[str, TrieNode] = {} # Mapping from char to TrieNode lowerCamelCase__ : List[str] = False def lowerCamelCase_ ( self: str , UpperCamelCase__: list[str] ): for word in words: self.insert(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : List[Any] = self for char in word: if char not in curr.nodes: lowerCamelCase__ : Tuple = TrieNode() lowerCamelCase__ : List[Any] = curr.nodes[char] lowerCamelCase__ : Any = True def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = self for char in word: if char not in curr.nodes: return False lowerCamelCase__ : Any = curr.nodes[char] return curr.is_leaf def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): def _delete(UpperCamelCase__: TrieNode , UpperCamelCase__: str , UpperCamelCase__: int ) -> bool: if index == len(UpperCamelCase__ ): # If word does not exist if not curr.is_leaf: return False lowerCamelCase__ : str = False return len(curr.nodes ) == 0 lowerCamelCase__ : List[str] = word[index] lowerCamelCase__ : Dict = curr.nodes.get(UpperCamelCase__ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCamelCase__ : List[Any] = _delete(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCamelCase__ , 0 ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: if node.is_leaf: print(UpperCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(UpperCamelCase , word + key ) def SCREAMING_SNAKE_CASE_ () -> bool: lowerCamelCase__ : str = """banana bananas bandana band apple all beast""".split() lowerCamelCase__ : Union[str, Any] = TrieNode() root.insert_many(UpperCamelCase ) # print_words(root, "") assert all(root.find(UpperCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: print(str(UpperCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def SCREAMING_SNAKE_CASE_ () -> None: assert test_trie() def SCREAMING_SNAKE_CASE_ () -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
41
1
'''simple docstring''' from manim import * class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[Any] = Rectangle(height=0.5 , width=0.5 ) lowerCamelCase__ : int = Rectangle(height=0.25 , width=0.25 ) lowerCamelCase__ : Optional[int] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) lowerCamelCase__ : List[str] = [mem.copy() for i in range(6 )] lowerCamelCase__ : List[Any] = [mem.copy() for i in range(6 )] lowerCamelCase__ : int = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : Dict = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : str = VGroup(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : List[str] = Text("""CPU""" , font_size=24 ) lowerCamelCase__ : Dict = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(UpperCamelCase__ ) lowerCamelCase__ : str = [mem.copy() for i in range(4 )] lowerCamelCase__ : Dict = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : Dict = Text("""GPU""" , font_size=24 ) lowerCamelCase__ : Union[str, Any] = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) gpu.move_to([-1, -1, 0] ) self.add(UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = [mem.copy() for i in range(6 )] lowerCamelCase__ : Tuple = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : List[Any] = Text("""Model""" , font_size=24 ) lowerCamelCase__ : Any = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) model.move_to([3, -1.0, 0] ) self.add(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = [] lowerCamelCase__ : Tuple = [] lowerCamelCase__ : Union[str, Any] = [] for i, rect in enumerate(UpperCamelCase__ ): rect.set_stroke(UpperCamelCase__ ) lowerCamelCase__ : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(UpperCamelCase__ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=UpperCamelCase__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=UpperCamelCase__ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=UpperCamelCase__ , buff=0.0 ) self.add(UpperCamelCase__ ) model_cpu_arr.append(UpperCamelCase__ ) self.add(*UpperCamelCase__ , *UpperCamelCase__ , *UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = [mem.copy() for i in range(6 )] lowerCamelCase__ : Any = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : Dict = Text("""Loaded Checkpoint""" , font_size=24 ) lowerCamelCase__ : Tuple = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) checkpoint.move_to([3, 0.5, 0] ) self.add(UpperCamelCase__ ) lowerCamelCase__ : Dict = [] lowerCamelCase__ : Dict = [] for i, rect in enumerate(UpperCamelCase__ ): lowerCamelCase__ : Tuple = fill.copy().set_fill(UpperCamelCase__ , opacity=0.7 ) target.move_to(UpperCamelCase__ ) ckpt_arr.append(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(UpperCamelCase__ ) self.add(*UpperCamelCase__ , *UpperCamelCase__ ) lowerCamelCase__ : str = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCamelCase__ : int = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(UpperCamelCase__ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(UpperCamelCase__ ) lowerCamelCase__ : Tuple = MarkupText( F'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) lowerCamelCase__ : int = [meta_mem.copy() for i in range(6 )] lowerCamelCase__ : Union[str, Any] = [meta_mem.copy() for i in range(6 )] lowerCamelCase__ : List[str] = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : Optional[int] = VGroup(*UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : List[str] = VGroup(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0 ) lowerCamelCase__ : Any = Text("""Disk""" , font_size=24 ) lowerCamelCase__ : List[str] = Group(UpperCamelCase__ , UpperCamelCase__ ).arrange(UpperCamelCase__ , buff=0.5 , aligned_edge=UpperCamelCase__ ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(UpperCamelCase__ , run_time=3 ) , Write(UpperCamelCase__ , run_time=1 ) , Create(UpperCamelCase__ , run_time=1 ) ) lowerCamelCase__ : Union[str, Any] = [] for i, rect in enumerate(UpperCamelCase__ ): lowerCamelCase__ : List[Any] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(UpperCamelCase__ , run_time=1.5 ) ) self.play(*UpperCamelCase__ ) self.play(FadeOut(UpperCamelCase__ ) ) lowerCamelCase__ : Any = MarkupText(F'''Then, the checkpoint is removed from memory\nthrough garbage collection.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCamelCase__ , run_time=3 ) ) self.play( FadeOut(UpperCamelCase__ , UpperCamelCase__ , *UpperCamelCase__ , *UpperCamelCase__ ) , ) self.wait()
41
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : Tuple = SwinConfig(image_size=192 ) if "base" in model_name: lowerCamelCase__ : Union[str, Any] = 6 lowerCamelCase__ : Dict = 128 lowerCamelCase__ : Any = (2, 2, 18, 2) lowerCamelCase__ : Tuple = (4, 8, 16, 32) elif "large" in model_name: lowerCamelCase__ : Any = 12 lowerCamelCase__ : Tuple = 192 lowerCamelCase__ : List[str] = (2, 2, 18, 2) lowerCamelCase__ : Union[str, Any] = (6, 12, 24, 48) else: raise ValueError("""Model not supported, only supports base and large variants""" ) lowerCamelCase__ : str = window_size lowerCamelCase__ : Dict = embed_dim lowerCamelCase__ : Tuple = depths lowerCamelCase__ : str = num_heads return config def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if "encoder.mask_token" in name: lowerCamelCase__ : Any = name.replace("""encoder.mask_token""" , """embeddings.mask_token""" ) if "encoder.patch_embed.proj" in name: lowerCamelCase__ : List[Any] = name.replace("""encoder.patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "encoder.patch_embed.norm" in name: lowerCamelCase__ : str = name.replace("""encoder.patch_embed.norm""" , """embeddings.norm""" ) if "attn.proj" in name: lowerCamelCase__ : List[Any] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: lowerCamelCase__ : List[Any] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: lowerCamelCase__ : Tuple = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: lowerCamelCase__ : int = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: lowerCamelCase__ : Optional[Any] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase__ : Optional[Any] = name.replace("""mlp.fc2""" , """output.dense""" ) if name == "encoder.norm.weight": lowerCamelCase__ : Optional[int] = """layernorm.weight""" if name == "encoder.norm.bias": lowerCamelCase__ : str = """layernorm.bias""" if "decoder" in name: pass else: lowerCamelCase__ : Tuple = """swin.""" + name return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[int]: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Tuple = orig_state_dict.pop(UpperCamelCase ) if "attn_mask" in key: pass elif "qkv" in key: lowerCamelCase__ : Union[str, Any] = key.split(""".""" ) lowerCamelCase__ : Optional[Any] = int(key_split[2] ) lowerCamelCase__ : Any = int(key_split[4] ) lowerCamelCase__ : List[str] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCamelCase__ : Union[str, Any] = val[:dim, :] lowerCamelCase__ : Tuple = val[ dim : dim * 2, : ] lowerCamelCase__ : Union[str, Any] = val[-dim:, :] else: lowerCamelCase__ : Union[str, Any] = val[ :dim ] lowerCamelCase__ : int = val[ dim : dim * 2 ] lowerCamelCase__ : List[str] = val[ -dim: ] else: lowerCamelCase__ : Union[str, Any] = val return orig_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = torch.load(UpperCamelCase , map_location="""cpu""" )["""model"""] lowerCamelCase__ : List[str] = get_swin_config(UpperCamelCase ) lowerCamelCase__ : str = SwinForMaskedImageModeling(UpperCamelCase ) model.eval() lowerCamelCase__ : List[Any] = convert_state_dict(UpperCamelCase , UpperCamelCase ) model.load_state_dict(UpperCamelCase ) lowerCamelCase__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Dict = ViTImageProcessor(size={"""height""": 192, """width""": 192} ) lowerCamelCase__ : Dict = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Union[str, Any] = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) with torch.no_grad(): lowerCamelCase__ : List[Any] = model(**UpperCamelCase ).logits print(outputs.keys() ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase ) if push_to_hub: print(f'''Pushing model and image processor for {model_name} to hub''' ) model.push_to_hub(f'''microsoft/{model_name}''' ) image_processor.push_to_hub(f'''microsoft/{model_name}''' ) if __name__ == "__main__": _A : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''swin-base-simmim-window6-192''', type=str, choices=['''swin-base-simmim-window6-192''', '''swin-large-simmim-window12-192'''], help='''Name of the Swin SimMIM model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default='''/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth''', type=str, help='''Path to the original PyTorch checkpoint (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _A : int =parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' from __future__ import annotations from statistics import mean def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list[int]: lowerCamelCase__ : Any = [0] * no_of_processes lowerCamelCase__ : Optional[int] = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(UpperCamelCase ): lowerCamelCase__ : Optional[Any] = burst_time[i] lowerCamelCase__ : list[int] = [] lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : Tuple = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: lowerCamelCase__ : int = [] lowerCamelCase__ : Optional[int] = -1 for i in range(UpperCamelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(UpperCamelCase ) if len(UpperCamelCase ) > 0: lowerCamelCase__ : Union[str, Any] = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: lowerCamelCase__ : int = i total_time += burst_time[target_process] completed += 1 lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : List[str] = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list[int]: lowerCamelCase__ : Optional[int] = [0] * no_of_processes for i in range(UpperCamelCase ): lowerCamelCase__ : int = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print('''[TEST CASE 01]''') _A : int =4 _A : Any =[2, 5, 3, 7] _A : Dict =[0, 0, 0, 0] _A : int =calculate_waitingtime(arrival_time, burst_time, no_of_processes) _A : int =calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print('''PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time''') for i, process_id in enumerate(list(range(1, 5))): print( F'{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t' F'{waiting_time[i]}\t\t\t\t{turn_around_time[i]}' ) print(F'\nAverage waiting time = {mean(waiting_time):.5f}') print(F'Average turnaround time = {mean(turn_around_time):.5f}')
41
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _A : List[Any] =logging.get_logger(__name__) _A : Dict =['''model.decoder.embed_positions.weights'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: if "emb" in name: lowerCamelCase__ : Dict = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: lowerCamelCase__ : List[str] = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: lowerCamelCase__ : List[str] = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: lowerCamelCase__ : Optional[int] = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: lowerCamelCase__ : Dict = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: lowerCamelCase__ : Optional[Any] = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: lowerCamelCase__ : Optional[Any] = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: lowerCamelCase__ : int = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple[Dict, Dict]: lowerCamelCase__ : int = list(state_dict.keys() ) lowerCamelCase__ : Tuple = {} for key in keys: lowerCamelCase__ : Any = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj lowerCamelCase__ : Union[str, Any] = val[:hidden_size, :] lowerCamelCase__ : Any = val[hidden_size : 2 * hidden_size, :] lowerCamelCase__ : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: lowerCamelCase__ : str = val else: lowerCamelCase__ : Union[str, Any] = val return state_dict, enc_dec_proj_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values lowerCamelCase__ : int = 1024 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 16 elif checkpoint == "medium": lowerCamelCase__ : Any = 1536 lowerCamelCase__ : Union[str, Any] = 48 lowerCamelCase__ : Optional[int] = 24 elif checkpoint == "large": lowerCamelCase__ : Optional[Any] = 2048 lowerCamelCase__ : Dict = 48 lowerCamelCase__ : List[Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) lowerCamelCase__ : Any = MusicgenDecoderConfig( hidden_size=UpperCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase , num_attention_heads=UpperCamelCase , ) return config @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase="cpu" ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = MusicGen.get_pretrained(UpperCamelCase , device=UpperCamelCase ) lowerCamelCase__ : List[Any] = decoder_config_from_checkpoint(UpperCamelCase ) lowerCamelCase__ : Any = fairseq_model.lm.state_dict() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = rename_state_dict( UpperCamelCase , hidden_size=decoder_config.hidden_size ) lowerCamelCase__ : str = TaEncoderModel.from_pretrained("""t5-base""" ) lowerCamelCase__ : Tuple = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) lowerCamelCase__ : Optional[int] = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection lowerCamelCase__ , lowerCamelCase__ : List[str] = decoder.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model lowerCamelCase__ : Optional[Any] = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase , audio_encoder=UpperCamelCase , decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass lowerCamelCase__ : Dict = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) lowerCamelCase__ : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor lowerCamelCase__ : str = AutoTokenizer.from_pretrained("""t5-base""" ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) lowerCamelCase__ : Optional[int] = MusicgenProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids lowerCamelCase__ : Union[str, Any] = 2048 lowerCamelCase__ : List[str] = 2048 # set other default generation config params lowerCamelCase__ : Optional[Any] = int(30 * audio_encoder.config.frame_rate ) lowerCamelCase__ : Union[str, Any] = True lowerCamelCase__ : List[Any] = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _A : List[str] =parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> float: lowerCamelCase__ : Tuple = 0 while len(UpperCamelCase ) > 1: lowerCamelCase__ : Dict = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): lowerCamelCase__ : Dict = files.index(min(UpperCamelCase ) ) temp += files[min_index] files.pop(UpperCamelCase ) files.append(UpperCamelCase ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' _A : int =''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' _A : List[Any] =[{'''type''': '''code''', '''content''': INSTALL_CONTENT}] _A : Union[str, Any] ={ '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
1
'''simple docstring''' _A : Tuple ={str(digit): digit**5 for digit in range(10)} def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ () -> int: return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(UpperCamelCase ) ) if __name__ == "__main__": print(solution())
41
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=7 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[str]=18 , UpperCamelCase__: Union[str, Any]=30 , UpperCamelCase__: List[str]=400 , UpperCamelCase__: Any=True , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[Any]=False , UpperCamelCase__: Tuple=True , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , ): lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Optional[int] = min_resolution lowerCamelCase__ : Optional[Any] = max_resolution lowerCamelCase__ : Union[str, Any] = do_resize lowerCamelCase__ : Optional[Any] = size if size is not None else {"""height""": 18, """width""": 20} lowerCamelCase__ : Dict = do_thumbnail lowerCamelCase__ : Optional[int] = do_align_axis lowerCamelCase__ : Any = do_pad lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Union[str, Any] = image_mean lowerCamelCase__ : Union[str, Any] = image_std def lowerCamelCase_ ( self: str ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DonutImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = DonutImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_thumbnail""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_align_long_axis""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 20} ) lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order lowerCamelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"""height""": 84, """width""": 42} ) def lowerCamelCase_ ( self: List[str] ): pass @is_flaky() def lowerCamelCase_ ( self: Union[str, Any] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Optional[int] ): # Initialize image_processing lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Dict ): # Initialize image_processing lowerCamelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
41
1
'''simple docstring''' from typing import List import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: lowerCamelCase__ : Any = {key: len(UpperCamelCase ) for key, value in gen_kwargs.items() if isinstance(UpperCamelCase , UpperCamelCase )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( """Sharding is ambiguous for this dataset: """ + """we found several data sources lists of different lengths, and we don't know over which list we should parallelize:\n""" + """\n""".join(f'''\t- key {key} has length {length}''' for key, length in lists_lengths.items() ) + """\nTo fix this, check the 'gen_kwargs' and make sure to use lists only for data sources, """ + """and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.""" ) ) lowerCamelCase__ : Any = max(lists_lengths.values() , default=0 ) return max(1 , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[range]: lowerCamelCase__ : Optional[Any] = [] for group_idx in range(UpperCamelCase ): lowerCamelCase__ : Tuple = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break lowerCamelCase__ : List[str] = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 lowerCamelCase__ : Tuple = range(UpperCamelCase , start + num_shards_to_add ) shards_indices_per_group.append(UpperCamelCase ) return shards_indices_per_group def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> List[dict]: lowerCamelCase__ : Optional[int] = _number_of_shards_in_gen_kwargs(UpperCamelCase ) if num_shards == 1: return [dict(UpperCamelCase )] else: lowerCamelCase__ : Union[str, Any] = _distribute_shards(num_shards=UpperCamelCase , max_num_jobs=UpperCamelCase ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(UpperCamelCase , UpperCamelCase ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(UpperCamelCase ) ) ] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> dict: return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , UpperCamelCase ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> dict: lowerCamelCase__ : int = {len(UpperCamelCase ) for value in gen_kwargs.values() if isinstance(UpperCamelCase , UpperCamelCase )} lowerCamelCase__ : Dict = {} for size in list_sizes: lowerCamelCase__ : int = list(range(UpperCamelCase ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes lowerCamelCase__ : Union[str, Any] = dict(UpperCamelCase ) for key, value in shuffled_kwargs.items(): if isinstance(UpperCamelCase , UpperCamelCase ): lowerCamelCase__ : Dict = [value[i] for i in indices_per_size[len(UpperCamelCase )]] return shuffled_kwargs
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowerCamelCase__ : List[Any] = torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = pipe( image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowerCamelCase__ : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase__ : Tuple = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: if len(UpperCamelCase ) != len(UpperCamelCase ): raise ValueError("""String lengths must match!""" ) lowerCamelCase__ : int = 0 for chara, chara in zip(UpperCamelCase , UpperCamelCase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _A : Union[str, Any] =False class _lowercase ( unittest.TestCase ): pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowerCamelCase__ : List[Any] = torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = pipe( image=UpperCamelCase__ , generator=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowerCamelCase__ : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase__ : Tuple = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
41
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _A : str =logging.get_logger(__name__) _A : List[str] ={ '''microsoft/focalnet-tiny''': '''https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json''', } class _lowercase ( _lowercase , _lowercase ): a = """focalnet""" def __init__( self: Dict , UpperCamelCase__: List[str]=224 , UpperCamelCase__: str=4 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: List[Any]=96 , UpperCamelCase__: Dict=False , UpperCamelCase__: Union[str, Any]=[192, 384, 768, 768] , UpperCamelCase__: Union[str, Any]=[2, 2, 6, 2] , UpperCamelCase__: Any=[2, 2, 2, 2] , UpperCamelCase__: Union[str, Any]=[3, 3, 3, 3] , UpperCamelCase__: Dict="gelu" , UpperCamelCase__: Dict=4.0 , UpperCamelCase__: Optional[int]=0.0 , UpperCamelCase__: str=0.1 , UpperCamelCase__: Union[str, Any]=False , UpperCamelCase__: str=1e-4 , UpperCamelCase__: List[str]=False , UpperCamelCase__: str=False , UpperCamelCase__: List[str]=False , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: str=1e-5 , UpperCamelCase__: Union[str, Any]=32 , UpperCamelCase__: int=None , UpperCamelCase__: Optional[int]=None , **UpperCamelCase__: int , ): super().__init__(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = image_size lowerCamelCase__ : Dict = patch_size lowerCamelCase__ : int = num_channels lowerCamelCase__ : str = embed_dim lowerCamelCase__ : List[str] = use_conv_embed lowerCamelCase__ : Dict = hidden_sizes lowerCamelCase__ : List[Any] = depths lowerCamelCase__ : int = focal_levels lowerCamelCase__ : Dict = focal_windows lowerCamelCase__ : str = hidden_act lowerCamelCase__ : Optional[Any] = mlp_ratio lowerCamelCase__ : int = hidden_dropout_prob lowerCamelCase__ : Any = drop_path_rate lowerCamelCase__ : Dict = use_layerscale lowerCamelCase__ : Union[str, Any] = layerscale_value lowerCamelCase__ : Optional[Any] = use_post_layernorm lowerCamelCase__ : Tuple = use_post_layernorm_in_modulation lowerCamelCase__ : Optional[int] = normalize_modulator lowerCamelCase__ : Dict = initializer_range lowerCamelCase__ : Any = layer_norm_eps lowerCamelCase__ : Union[str, Any] = encoder_stride lowerCamelCase__ : Union[str, Any] = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] lowerCamelCase__ , lowerCamelCase__ : str = get_aligned_output_features_output_indices( out_features=UpperCamelCase__ , out_indices=UpperCamelCase__ , stage_names=self.stage_names )
41
'''simple docstring''' # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _A : Dict ='''tiny-wmt19-en-ru''' # Build # borrowed from a test _A : List[str] =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _A : str =dict(zip(vocab, range(len(vocab)))) _A : List[str] =['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: _A : Union[str, Any] =Path(tmpdirname) _A : str =build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] _A : int =build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] _A : List[Any] =build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) _A : int =FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _A : List[str] =FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _A : Union[str, Any] =FSMTForConditionalGeneration(config) print(F'num of params {tiny_model.num_parameters()}') # Test _A : List[str] =tokenizer(['''Making tiny model'''], return_tensors='''pt''') _A : Tuple =tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'Generated {mname_tiny}') # Upload # transformers-cli upload tiny-wmt19-en-ru
41
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _A : Any ={'''configuration_unispeech''': ['''UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''UniSpeechConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Any =[ '''UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST''', '''UniSpeechForCTC''', '''UniSpeechForPreTraining''', '''UniSpeechForSequenceClassification''', '''UniSpeechModel''', '''UniSpeechPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys _A : Dict =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
'''simple docstring''' from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Tuple=4 , UpperCamelCase__: Optional[Any]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: Union[str, Any]=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: str=0.6 , UpperCamelCase__: str=None , ): lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Optional[Any] = batch_size lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Optional[Any] = is_training lowerCamelCase__ : int = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[Any] = hidden_act lowerCamelCase__ : Any = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : Optional[int] = initializer_range lowerCamelCase__ : List[str] = mask_ratio lowerCamelCase__ : Optional[int] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Any = (image_size // patch_size) ** 2 lowerCamelCase__ : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Any ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel(config=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Tuple , UpperCamelCase__: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , training=UpperCamelCase__ ) # expected sequence length = num_patches lowerCamelCase__ : List[str] = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Union[str, Any] = TFViTMAEForPreTraining(UpperCamelCase__ ) lowerCamelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase__ : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Optional[int] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () a = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : int = TFViTMAEModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : int = copy.deepcopy(self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : str = outputs_dict[0].numpy() lowerCamelCase__ : Optional[int] = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def lowerCamelCase_ ( self: Dict ): # make the mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(UpperCamelCase__: int ): lowerCamelCase__ : Optional[int] = {} for k, v in inputs_dict.items(): if tf.is_tensor(UpperCamelCase__ ): lowerCamelCase__ : List[str] = v.numpy() else: lowerCamelCase__ : Union[str, Any] = np.array(UpperCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : str = prepare_numpy_arrays(UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : Any = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: str ): # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase__ : int = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : Tuple = tf_noise super().check_pt_tf_models(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[Any] = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(UpperCamelCase__ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(UpperCamelCase__ , UpperCamelCase__ ),) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(UpperCamelCase__ , """_keras_serializable""" , UpperCamelCase__ ) } lowerCamelCase__ : List[str] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCamelCase__ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase__ : List[str] = main_layer_class(UpperCamelCase__ ) lowerCamelCase__ : int = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase__ : List[str] = tf.keras.Model(UpperCamelCase__ , outputs=main_layer(UpperCamelCase__ ) ) lowerCamelCase__ : Union[str, Any] = model(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : int = os.path.join(UpperCamelCase__ , """keras_model.h5""" ) model.save(UpperCamelCase__ ) lowerCamelCase__ : int = tf.keras.models.load_model( UpperCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(UpperCamelCase__ , tf.keras.Model ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: str ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : Any = outputs.last_hidden_state.numpy() lowerCamelCase__ : List[str] = 0 else: lowerCamelCase__ : int = outputs.logits.numpy() lowerCamelCase__ : Dict = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCamelCase__ , saved_model=UpperCamelCase__ ) lowerCamelCase__ : Dict = model_class.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ , noise=UpperCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase__ : str = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase__ : Optional[Any] = 0 else: lowerCamelCase__ : Union[str, Any] = after_outputs["""logits"""].numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCamelCase__ , 1e-5 ) def lowerCamelCase_ ( self: Any ): # make mask reproducible np.random.seed(2 ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Optional[int] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase__ : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(UpperCamelCase__ , noise=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(UpperCamelCase__ ) lowerCamelCase__ : int = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase__ : int = model_class.from_config(model.config ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase__ : List[Any] = new_model(UpperCamelCase__ , noise=UpperCamelCase__ ) self.assert_outputs_same(UpperCamelCase__ , UpperCamelCase__ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @slow def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Tuple = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[Any] ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self: List[str] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : Optional[int] = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Tuple = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : str = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase__ : str = model(**UpperCamelCase__ , noise=UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Any = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , UpperCamelCase__ , atol=1e-4 )
41
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A : Union[str, Any] =logging.get_logger(__name__) _A : Dict ={ '''hustvl/yolos-small''': '''https://huggingface.co/hustvl/yolos-small/resolve/main/config.json''', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowercase ( _lowercase ): a = """yolos""" def __init__( self: List[Any] , UpperCamelCase__: Any=768 , UpperCamelCase__: Optional[Any]=12 , UpperCamelCase__: int=12 , UpperCamelCase__: int=3_072 , UpperCamelCase__: Dict="gelu" , UpperCamelCase__: int=0.0 , UpperCamelCase__: Dict=0.0 , UpperCamelCase__: int=0.02 , UpperCamelCase__: Optional[int]=1e-12 , UpperCamelCase__: Optional[Any]=[512, 864] , UpperCamelCase__: List[Any]=16 , UpperCamelCase__: str=3 , UpperCamelCase__: Dict=True , UpperCamelCase__: Optional[int]=100 , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Tuple=False , UpperCamelCase__: int=1 , UpperCamelCase__: str=5 , UpperCamelCase__: List[str]=2 , UpperCamelCase__: Optional[Any]=5 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: Dict=0.1 , **UpperCamelCase__: List[Any] , ): super().__init__(**UpperCamelCase__ ) lowerCamelCase__ : Any = hidden_size lowerCamelCase__ : str = num_hidden_layers lowerCamelCase__ : List[str] = num_attention_heads lowerCamelCase__ : int = intermediate_size lowerCamelCase__ : List[Any] = hidden_act lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Dict = layer_norm_eps lowerCamelCase__ : List[str] = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Dict = qkv_bias lowerCamelCase__ : Optional[int] = num_detection_tokens lowerCamelCase__ : Optional[int] = use_mid_position_embeddings lowerCamelCase__ : Union[str, Any] = auxiliary_loss # Hungarian matcher lowerCamelCase__ : int = class_cost lowerCamelCase__ : Any = bbox_cost lowerCamelCase__ : str = giou_cost # Loss coefficients lowerCamelCase__ : Dict = bbox_loss_coefficient lowerCamelCase__ : Tuple = giou_loss_coefficient lowerCamelCase__ : int = eos_coefficient class _lowercase ( _lowercase ): a = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self: Union[str, Any] ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCamelCase_ ( self: Tuple ): return 1e-4 @property def lowerCamelCase_ ( self: str ): return 12
41
'''simple docstring''' class _lowercase : def __init__( self: Tuple , UpperCamelCase__: list[int] ): lowerCamelCase__ : Union[str, Any] = len(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = [0] * len_array if len_array > 0: lowerCamelCase__ : Union[str, Any] = array[0] for i in range(1 , UpperCamelCase__ ): lowerCamelCase__ : Tuple = self.prefix_sum[i - 1] + array[i] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: int ): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: int ): lowerCamelCase__ : Dict = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(UpperCamelCase__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar _A : str =TypeVar('''KEY''') _A : Optional[Any] =TypeVar('''VAL''') @dataclass(frozen=_lowercase , slots=_lowercase ) class _lowercase ( Generic[KEY, VAL] ): a = 42 a = 42 class _lowercase ( _Item ): def __init__( self: List[str] ): super().__init__(UpperCamelCase__ , UpperCamelCase__ ) def __bool__( self: Optional[Any] ): return False _A : List[str] =_DeletedItem() class _lowercase ( MutableMapping[KEY, VAL] ): def __init__( self: Any , UpperCamelCase__: int = 8 , UpperCamelCase__: float = 0.75 ): lowerCamelCase__ : Optional[Any] = initial_block_size lowerCamelCase__ : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 lowerCamelCase__ : Union[str, Any] = capacity_factor lowerCamelCase__ : List[Any] = 0 def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: KEY ): return hash(UpperCamelCase__ ) % len(self._buckets ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: int ): return (ind + 1) % len(self._buckets ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: int , UpperCamelCase__: KEY , UpperCamelCase__: VAL ): lowerCamelCase__ : Any = self._buckets[ind] if not stored: lowerCamelCase__ : Optional[int] = _Item(UpperCamelCase__ , UpperCamelCase__ ) self._len += 1 return True elif stored.key == key: lowerCamelCase__ : Tuple = _Item(UpperCamelCase__ , UpperCamelCase__ ) return True else: return False def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : List[Any] = len(self._buckets ) * self._capacity_factor return len(self ) >= int(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): if len(self._buckets ) <= self._initial_block_size: return False lowerCamelCase__ : Optional[int] = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: int ): lowerCamelCase__ : Tuple = self._buckets lowerCamelCase__ : Union[str, Any] = [None] * new_size lowerCamelCase__ : List[Any] = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def lowerCamelCase_ ( self: Optional[Any] ): self._resize(len(self._buckets ) * 2 ) def lowerCamelCase_ ( self: Union[str, Any] ): self._resize(len(self._buckets ) // 2 ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: KEY ): lowerCamelCase__ : Tuple = self._get_bucket_index(UpperCamelCase__ ) for _ in range(len(self._buckets ) ): yield ind lowerCamelCase__ : int = self._get_next_ind(UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: KEY , UpperCamelCase__: VAL ): for ind in self._iterate_buckets(UpperCamelCase__ ): if self._try_set(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): break def __setitem__( self: Optional[Any] , UpperCamelCase__: KEY , UpperCamelCase__: VAL ): if self._is_full(): self._size_up() self._add_item(UpperCamelCase__ , UpperCamelCase__ ) def __delitem__( self: Tuple , UpperCamelCase__: KEY ): for ind in self._iterate_buckets(UpperCamelCase__ ): lowerCamelCase__ : List[Any] = self._buckets[ind] if item is None: raise KeyError(UpperCamelCase__ ) if item is _deleted: continue if item.key == key: lowerCamelCase__ : Union[str, Any] = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self: Dict , UpperCamelCase__: KEY ): for ind in self._iterate_buckets(UpperCamelCase__ ): lowerCamelCase__ : Optional[int] = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(UpperCamelCase__ ) def __len__( self: str ): return self._len def __iter__( self: Union[str, Any] ): yield from (item.key for item in self._buckets if item) def __repr__( self: List[str] ): lowerCamelCase__ : Tuple = """ ,""".join( F'''{item.key}: {item.val}''' for item in self._buckets if item ) return F'''HashMap({val_string})'''
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A : Dict ={ '''configuration_swinv2''': ['''SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Swinv2Config'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] =[ '''SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Swinv2ForImageClassification''', '''Swinv2ForMaskedImageModeling''', '''Swinv2Model''', '''Swinv2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: if digit_amount > 0: return round(number - int(UpperCamelCase ) , UpperCamelCase ) return number - int(UpperCamelCase ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
41
'''simple docstring''' from __future__ import annotations _A : Any ={ '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class _lowercase : def __init__( self: Tuple , UpperCamelCase__: dict[str, list[str]] , UpperCamelCase__: str ): lowerCamelCase__ : str = graph # mapping node to its parent in resulting breadth first tree lowerCamelCase__ : dict[str, str | None] = {} lowerCamelCase__ : Any = source_vertex def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = {self.source_vertex} lowerCamelCase__ : List[str] = None lowerCamelCase__ : Tuple = [self.source_vertex] # first in first out queue while queue: lowerCamelCase__ : Tuple = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCamelCase__ ) lowerCamelCase__ : List[str] = vertex queue.append(UpperCamelCase__ ) def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): if target_vertex == self.source_vertex: return self.source_vertex lowerCamelCase__ : Tuple = self.parent.get(UpperCamelCase__ ) if target_vertex_parent is None: lowerCamelCase__ : int = ( F'''No path from vertex: {self.source_vertex} to vertex: {target_vertex}''' ) raise ValueError(UpperCamelCase__ ) return self.shortest_path(UpperCamelCase__ ) + F'''->{target_vertex}''' if __name__ == "__main__": _A : int =Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
41
1
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class _lowercase ( _lowercase , _lowercase , _lowercase , unittest.TestCase ): a = StableDiffusionControlNetImgaImgPipeline a = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} a = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({"""control_image"""} ) a = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase_ ( self: Any ): torch.manual_seed(0 ) lowerCamelCase__ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=UpperCamelCase__ , set_alpha_to_one=UpperCamelCase__ , ) torch.manual_seed(0 ) lowerCamelCase__ : List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCamelCase__ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) lowerCamelCase__ : str = CLIPTextModel(UpperCamelCase__ ) lowerCamelCase__ : Any = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCamelCase__ : Dict = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCamelCase_ ( self: str , UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple=0 ): if str(UpperCamelCase__ ).startswith("""mps""" ): lowerCamelCase__ : Any = torch.manual_seed(UpperCamelCase__ ) else: lowerCamelCase__ : int = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) lowerCamelCase__ : Tuple = 2 lowerCamelCase__ : Optional[Any] = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=UpperCamelCase__ , device=torch.device(UpperCamelCase__ ) , ) lowerCamelCase__ : Optional[Any] = floats_tensor(control_image.shape , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) lowerCamelCase__ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase__ : List[str] = Image.fromarray(np.uinta(UpperCamelCase__ ) ).convert("""RGB""" ).resize((64, 64) ) lowerCamelCase__ : Tuple = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowerCamelCase_ ( self: Union[str, Any] ): return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowerCamelCase_ ( self: Optional[int] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def lowerCamelCase_ ( self: Any ): self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = StableDiffusionControlNetImgaImgPipeline a = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} a = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowerCamelCase_ ( self: str ): torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(UpperCamelCase__: Tuple ): if isinstance(UpperCamelCase__ , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) lowerCamelCase__ : Optional[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(UpperCamelCase__ ) torch.manual_seed(0 ) lowerCamelCase__ : Tuple = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(UpperCamelCase__ ) torch.manual_seed(0 ) lowerCamelCase__ : Optional[Any] = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=UpperCamelCase__ , set_alpha_to_one=UpperCamelCase__ , ) torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCamelCase__ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) lowerCamelCase__ : Optional[int] = CLIPTextModel(UpperCamelCase__ ) lowerCamelCase__ : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCamelCase__ : Any = MultiControlNetModel([controlneta, controlneta] ) lowerCamelCase__ : List[Any] = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: str , UpperCamelCase__: Optional[Any]=0 ): if str(UpperCamelCase__ ).startswith("""mps""" ): lowerCamelCase__ : int = torch.manual_seed(UpperCamelCase__ ) else: lowerCamelCase__ : Optional[Any] = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) lowerCamelCase__ : Any = 2 lowerCamelCase__ : List[str] = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=UpperCamelCase__ , device=torch.device(UpperCamelCase__ ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=UpperCamelCase__ , device=torch.device(UpperCamelCase__ ) , ), ] lowerCamelCase__ : Tuple = floats_tensor(control_image[0].shape , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase__ : Any = Image.fromarray(np.uinta(UpperCamelCase__ ) ).convert("""RGB""" ).resize((64, 64) ) lowerCamelCase__ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.get_dummy_components() lowerCamelCase__ : List[str] = self.pipeline_class(**UpperCamelCase__ ) pipe.to(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = 10.0 lowerCamelCase__ : Any = 4 lowerCamelCase__ : Union[str, Any] = self.get_dummy_inputs(UpperCamelCase__ ) lowerCamelCase__ : List[str] = steps lowerCamelCase__ : Optional[int] = scale lowerCamelCase__ : Union[str, Any] = pipe(**UpperCamelCase__ )[0] lowerCamelCase__ : Any = self.get_dummy_inputs(UpperCamelCase__ ) lowerCamelCase__ : Any = steps lowerCamelCase__ : Union[str, Any] = scale lowerCamelCase__ : Tuple = pipe(**UpperCamelCase__ , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] lowerCamelCase__ : str = self.get_dummy_inputs(UpperCamelCase__ ) lowerCamelCase__ : Tuple = steps lowerCamelCase__ : int = scale lowerCamelCase__ : List[str] = pipe(**UpperCamelCase__ , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] lowerCamelCase__ : Optional[Any] = self.get_dummy_inputs(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = steps lowerCamelCase__ : Optional[Any] = scale lowerCamelCase__ : Optional[Any] = pipe(**UpperCamelCase__ , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 def lowerCamelCase_ ( self: str ): return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowerCamelCase_ ( self: int ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def lowerCamelCase_ ( self: int ): self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[Any] = self.get_dummy_components() lowerCamelCase__ : str = self.pipeline_class(**UpperCamelCase__ ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(UpperCamelCase__ ) except NotImplementedError: pass @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: List[str] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Optional[int] = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) lowerCamelCase__ : Tuple = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=UpperCamelCase__ , controlnet=UpperCamelCase__ ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowerCamelCase__ : List[Any] = """evil space-punk bird""" lowerCamelCase__ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((512, 512) ) lowerCamelCase__ : Union[str, Any] = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((512, 512) ) lowerCamelCase__ : Optional[int] = pipe( UpperCamelCase__ , UpperCamelCase__ , control_image=UpperCamelCase__ , generator=UpperCamelCase__ , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) lowerCamelCase__ : Optional[int] = output.images[0] assert image.shape == (512, 512, 3) lowerCamelCase__ : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9e-2
41
'''simple docstring''' from collections.abc import Sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(UpperCamelCase ) ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> float: lowerCamelCase__ : str = 0.0 for coeff in reversed(UpperCamelCase ): lowerCamelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": _A : Any =(0.0, 0.0, 5.0, 9.3, 7.0) _A : Optional[Any] =10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
41
1
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowercase : def __init__( self: List[str] , UpperCamelCase__: Dict , UpperCamelCase__: Dict=13 , UpperCamelCase__: int=30 , UpperCamelCase__: Tuple=2 , UpperCamelCase__: Any=3 , UpperCamelCase__: List[str]=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple=5 , UpperCamelCase__: Any=4 , UpperCamelCase__: str=37 , UpperCamelCase__: Union[str, Any]="gelu" , UpperCamelCase__: str=0.1 , UpperCamelCase__: List[Any]=0.1 , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: Optional[Any]=3 , UpperCamelCase__: Any=None , UpperCamelCase__: Any=2 , ): lowerCamelCase__ : str = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : Any = patch_size lowerCamelCase__ : Dict = num_channels lowerCamelCase__ : List[str] = is_training lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : str = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = intermediate_size lowerCamelCase__ : str = hidden_act lowerCamelCase__ : int = hidden_dropout_prob lowerCamelCase__ : Dict = attention_probs_dropout_prob lowerCamelCase__ : int = type_sequence_label_size lowerCamelCase__ : Tuple = initializer_range lowerCamelCase__ : Optional[int] = scope lowerCamelCase__ : Tuple = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) lowerCamelCase__ : str = (image_size // patch_size) ** 2 lowerCamelCase__ : int = num_patches + 2 def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[int] = None if self.use_labels: lowerCamelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[int] = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self: Dict ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): lowerCamelCase__ : Any = DeiTModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Tuple ): lowerCamelCase__ : List[str] = DeiTForMaskedImageModeling(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Any = model(UpperCamelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Optional[int] = DeiTForMaskedImageModeling(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : int = model(UpperCamelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Dict , UpperCamelCase__: Tuple , UpperCamelCase__: List[str] ): lowerCamelCase__ : str = self.type_sequence_label_size lowerCamelCase__ : Union[str, Any] = DeiTForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : List[Any] = DeiTForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[str] = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Union[str, Any] = config_and_inputs lowerCamelCase__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) a = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) a = False a = False a = False def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[str] = DeiTModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: int ): self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCamelCase__ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__ , nn.Linear ) ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[Any] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Tuple = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[str] , UpperCamelCase__: Dict=False ): lowerCamelCase__ : List[Any] = super()._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCamelCase_ ( self: Dict ): if not self.model_tester.is_training: return lowerCamelCase__ , lowerCamelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : List[str] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(UpperCamelCase__ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() lowerCamelCase__ : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase__ : str = model(**UpperCamelCase__ ).loss loss.backward() def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowerCamelCase__ : Tuple = False lowerCamelCase__ : Any = True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue lowerCamelCase__ : Optional[int] = model_class(UpperCamelCase__ ) model.gradient_checkpointing_enable() model.to(UpperCamelCase__ ) model.train() lowerCamelCase__ : Optional[int] = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase__ : int = model(**UpperCamelCase__ ).loss loss.backward() def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : str = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(UpperCamelCase__ ), *get_values(UpperCamelCase__ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type['title']}''' ): lowerCamelCase__ : int = problem_type["""title"""] lowerCamelCase__ : Any = problem_type["""num_labels"""] lowerCamelCase__ : List[Any] = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() lowerCamelCase__ : Dict = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) if problem_type["num_labels"] > 1: lowerCamelCase__ : List[Any] = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) lowerCamelCase__ : List[Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=UpperCamelCase__ ) as warning_list: lowerCamelCase__ : List[Any] = model(**UpperCamelCase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def lowerCamelCase_ ( self: Union[str, Any] ): for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Optional[Any] = DeiTModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> List[str]: lowerCamelCase__ : List[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: List[Any] ): return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[int] = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : Optional[int] = prepare_img() lowerCamelCase__ : str = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : List[str] = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : Union[str, Any] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : Any = torch.tensor([-1.0_266, 0.1_912, -1.2_861] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) lowerCamelCase__ : List[Any] = self.default_image_processor lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = inputs.pixel_values.to(UpperCamelCase__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCamelCase__ : int = model(UpperCamelCase__ )
41
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): _A : List[Any] ='''pt''' elif is_tf_available(): _A : Any ='''tf''' else: _A : List[str] ='''jax''' class _lowercase ( _lowercase , unittest.TestCase ): a = ByTaTokenizer a = False def lowerCamelCase_ ( self: str ): super().setUp() lowerCamelCase__ : str = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ByTaTokenizer.from_pretrained("""google/byt5-small""" ) def lowerCamelCase_ ( self: Any , **UpperCamelCase__: Tuple ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Any=False , UpperCamelCase__: Union[str, Any]=20 , UpperCamelCase__: Optional[int]=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase__ ) ): try: lowerCamelCase__ : Tuple = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase__ ) except UnicodeDecodeError: pass toks.append((i, tok) ) lowerCamelCase__ : Union[str, Any] = list(filter(lambda UpperCamelCase__ : re.match(R"""^[ a-zA-Z]+$""" , t[1] ) , UpperCamelCase__ ) ) lowerCamelCase__ : Tuple = list(filter(lambda UpperCamelCase__ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase__ ) , UpperCamelCase__ ) ) if max_length is not None and len(UpperCamelCase__ ) > max_length: lowerCamelCase__ : Dict = toks[:max_length] if min_length is not None and len(UpperCamelCase__ ) < min_length and len(UpperCamelCase__ ) > 0: while len(UpperCamelCase__ ) < min_length: lowerCamelCase__ : Optional[Any] = toks + toks # toks_str = [t[1] for t in toks] lowerCamelCase__ : Union[str, Any] = [t[0] for t in toks] # Ensure consistency lowerCamelCase__ : Union[str, Any] = tokenizer.decode(UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) if " " not in output_txt and len(UpperCamelCase__ ) > 1: lowerCamelCase__ : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase__ ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase__ ) ) if with_prefix_space: lowerCamelCase__ : str = """ """ + output_txt lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) return output_txt, output_ids def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : Union[str, Any] = tokenizer(["""hi</s>""", """I went to the gym</s>""", """</s>"""] ) lowerCamelCase__ : Optional[int] = tokenizer(["""hi""", """I went to the gym""", """"""] ) self.assertListEqual(batch_with_eos_added["""input_ids"""] , batch_without_eos_added["""input_ids"""] ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[Any] = self.ta_base_tokenizer lowerCamelCase__ : Dict = """Unicode €.""" lowerCamelCase__ : List[Any] = tokenizer(UpperCamelCase__ ) lowerCamelCase__ : List[str] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : Tuple = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """Unicode €.</s>""" ) lowerCamelCase__ : List[Any] = tokenizer("""e è é ê ë""" ) lowerCamelCase__ : List[Any] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["""input_ids"""] , UpperCamelCase__ ) # decoding lowerCamelCase__ : str = tokenizer.decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , """e è é ê ë</s>""" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("""e è é ê ë""" ) ) , """e è é ê ë</s>""" ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : int = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off lowerCamelCase__ : List[str] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on lowerCamelCase__ : int = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) if FRAMEWORK != "jax": lowerCamelCase__ : Any = list(batch.input_ids.numpy()[0] ) else: lowerCamelCase__ : str = list(batch.input_ids.tolist()[0] ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : List[str] = self.ta_base_tokenizer lowerCamelCase__ : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase__ : Tuple = tokenizer(UpperCamelCase__ , padding=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("""input_ids""" , UpperCamelCase__ ) self.assertIn("""attention_mask""" , UpperCamelCase__ ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase__ ) self.assertNotIn("""decoder_attention_mask""" , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : str = self.ta_base_tokenizer lowerCamelCase__ : List[Any] = [ """Summary of the text.""", """Another summary.""", ] lowerCamelCase__ : Union[str, Any] = tokenizer( text_target=UpperCamelCase__ , max_length=32 , padding="""max_length""" , truncation=UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.ta_base_tokenizer lowerCamelCase__ : str = ["""A long paragraph for summarization. </s>"""] lowerCamelCase__ : Optional[Any] = ["""Summary of the text. </s>"""] # fmt: off lowerCamelCase__ : Tuple = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] lowerCamelCase__ : Any = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ , text_target=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , batch["""input_ids"""][0] ) self.assertEqual(UpperCamelCase__ , batch["""labels"""][0] ) def lowerCamelCase_ ( self: Optional[int] ): # safety check on max_len default value so we are sure the test works lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test lowerCamelCase__ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : int = tempfile.mkdtemp() lowerCamelCase__ : List[str] = """ He is very happy, UNwant\u00E9d,running""" lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) shutil.rmtree(UpperCamelCase__ ) lowerCamelCase__ : Any = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc lowerCamelCase__ : Any = tempfile.mkdtemp() lowerCamelCase__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(["""bim""", """bambam"""] ) lowerCamelCase__ : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("""new_additional_special_token""" ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) lowerCamelCase__ : List[Any] = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : int = after_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertIn("""new_additional_special_token""" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) lowerCamelCase__ : Any = tokenizer.__class__.from_pretrained(UpperCamelCase__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Union[str, Any] = json.load(UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: lowerCamelCase__ : Optional[Any] = json.load(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = [F'''<extra_id_{i}>''' for i in range(125 )] lowerCamelCase__ : int = added_tokens_extra_ids + [ """an_additional_special_token""" ] lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(UpperCamelCase__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) with open(os.path.join(UpperCamelCase__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files lowerCamelCase__ : Dict = tokenizer_class.from_pretrained( UpperCamelCase__ , ) self.assertIn( """an_additional_special_token""" , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["""an_additional_special_token"""] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["""an_additional_special_token"""] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained lowerCamelCase__ : Optional[Any] = added_tokens_extra_ids + [AddedToken("""a_new_additional_special_token""" , lstrip=UpperCamelCase__ )] lowerCamelCase__ : Any = tokenizer_class.from_pretrained( UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , ) self.assertIn("""a_new_additional_special_token""" , tokenizer.additional_special_tokens ) self.assertEqual( ["""a_new_additional_special_token"""] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["""a_new_additional_special_token"""] ) ) , ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase__ ) lowerCamelCase__ : str = tokenizer_class.from_pretrained(UpperCamelCase__ ) self.assertTrue(tokenizer.decode([255] ) == """""" ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: str ): pass def lowerCamelCase_ ( self: List[str] ): pass def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: int ): # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens lowerCamelCase__ : Dict = self.get_tokenizers(fast=UpperCamelCase__ , do_lower_case=UpperCamelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : Dict = ["""t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """x""", """t""", """</s>"""] lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowerCamelCase__ : str = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] lowerCamelCase__ : str = 0 lowerCamelCase__ : Any = tokenizer.convert_ids_to_tokens( UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) for attr in attributes_list: setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , attr + """_id""" , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(getattr(UpperCamelCase__ , attr + """_id""" ) , UpperCamelCase__ ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [] ) setattr(UpperCamelCase__ , """additional_special_tokens_ids""" , [token_id_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens""" ) , [token_to_test_setters] ) self.assertListEqual(getattr(UpperCamelCase__ , """additional_special_tokens_ids""" ) , [token_id_to_test_setters] )
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: lowerCamelCase__ : Dict = int(UpperCamelCase ) if decimal in (0, 1): # Exit cases for the recursion return str(UpperCamelCase ) lowerCamelCase__ , lowerCamelCase__ : str = divmod(UpperCamelCase , 2 ) return binary_recursive(UpperCamelCase ) + str(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: lowerCamelCase__ : int = str(UpperCamelCase ).strip() if not number: raise ValueError("""No input value was provided""" ) lowerCamelCase__ : Dict = """-""" if number.startswith("""-""" ) else """""" lowerCamelCase__ : List[str] = number.lstrip("""-""" ) if not number.isnumeric(): raise ValueError("""Input value is not an integer""" ) return f'''{negative}0b{binary_recursive(int(UpperCamelCase ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
41
'''simple docstring''' from heapq import heappop, heappush import numpy as np def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) -> tuple[float | int, list[tuple[int, int]]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = grid.shape lowerCamelCase__ : List[str] = [-1, 1, 0, 0] lowerCamelCase__ : Dict = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] lowerCamelCase__ , lowerCamelCase__ : Any = [(0, source)], set() lowerCamelCase__ : Tuple = np.full((rows, cols) , np.inf ) lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[int] = np.empty((rows, cols) , dtype=UpperCamelCase ) lowerCamelCase__ : str = None while queue: ((lowerCamelCase__) , (lowerCamelCase__)) : List[str] = heappop(UpperCamelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: lowerCamelCase__ : Optional[int] = [] while (x, y) != source: path.append((x, y) ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = predecessors[x, y] path.append(UpperCamelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCamelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: lowerCamelCase__ : Any = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCamelCase , (dist + 1, (nx, ny)) ) lowerCamelCase__ : Union[str, Any] = dist + 1 lowerCamelCase__ : List[str] = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
41
1
'''simple docstring''' from __future__ import annotations _A : List[Any] ='''Muhammad Umer Farooq''' _A : Union[str, Any] ='''MIT''' _A : List[Any] ='''1.0.0''' _A : List[str] ='''Muhammad Umer Farooq''' _A : str ='''contact@muhammadumerfarooq.me''' _A : List[Any] ='''Alpha''' import re from html.parser import HTMLParser from urllib import parse import requests class _lowercase ( _lowercase ): def __init__( self: Tuple , UpperCamelCase__: str ): super().__init__() lowerCamelCase__ : list[str] = [] lowerCamelCase__ : int = domain def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str , UpperCamelCase__: list[tuple[str, str | None]] ): # Only parse the 'anchor' tag. if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: lowerCamelCase__ : str = parse.urljoin(self.domain , UpperCamelCase__ ) self.urls.append(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: return ".".join(get_sub_domain_name(UpperCamelCase ).split(""".""" )[-2:] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: return parse.urlparse(UpperCamelCase ).netloc def SCREAMING_SNAKE_CASE_ (UpperCamelCase = "https://github.com" ) -> list[str]: lowerCamelCase__ : List[Any] = get_domain_name(UpperCamelCase ) # Initialize the parser lowerCamelCase__ : Optional[int] = Parser(UpperCamelCase ) try: # Open URL lowerCamelCase__ : Union[str, Any] = requests.get(UpperCamelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through lowerCamelCase__ : List[str] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: lowerCamelCase__ : Optional[Any] = requests.get(UpperCamelCase ) # Get the valid email. lowerCamelCase__ : Union[str, Any] = re.findall("""[a-zA-Z0-9]+@""" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(UpperCamelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(UpperCamelCase ) if __name__ == "__main__": _A : List[str] =emails_from_url('''https://github.com''') print(F'{len(emails)} emails found:') print('''\n'''.join(sorted(emails)))
41
'''simple docstring''' from __future__ import annotations import requests _A : str =set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase = 1 , UpperCamelCase = "new" , UpperCamelCase = None ) -> dict: lowerCamelCase__ : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(UpperCamelCase ) - valid_terms ) ): lowerCamelCase__ : str = f'''Invalid search term: {invalid_search_terms}''' raise ValueError(UpperCamelCase ) lowerCamelCase__ : str = requests.get( f'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''' , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : Optional[int] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(UpperCamelCase )} lowerCamelCase__ : Dict = {} for id_ in range(UpperCamelCase ): lowerCamelCase__ : Union[str, Any] = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
41
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class _lowercase : def __init__( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Optional[Any]=13 , UpperCamelCase__: Dict=7 , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=True , UpperCamelCase__: List[str]=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: List[str]=99 , UpperCamelCase__: Dict=32 , UpperCamelCase__: int=2 , UpperCamelCase__: Any=4 , UpperCamelCase__: Optional[int]=37 , UpperCamelCase__: List[Any]="gelu" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[str]=0.1 , UpperCamelCase__: Dict=512 , UpperCamelCase__: Union[str, Any]=16 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: int=0.02 , UpperCamelCase__: Tuple=3 , UpperCamelCase__: Optional[Any]=4 , UpperCamelCase__: Optional[int]=None , UpperCamelCase__: List[Any]=0 , ): lowerCamelCase__ : str = parent lowerCamelCase__ : Tuple = batch_size lowerCamelCase__ : Optional[int] = seq_length lowerCamelCase__ : Dict = is_training lowerCamelCase__ : Optional[int] = use_input_mask lowerCamelCase__ : Optional[Any] = use_token_type_ids lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Tuple = vocab_size lowerCamelCase__ : List[str] = hidden_size lowerCamelCase__ : List[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : Optional[int] = intermediate_size lowerCamelCase__ : List[str] = hidden_act lowerCamelCase__ : Dict = hidden_dropout_prob lowerCamelCase__ : Dict = attention_probs_dropout_prob lowerCamelCase__ : str = max_position_embeddings lowerCamelCase__ : List[str] = type_vocab_size lowerCamelCase__ : Optional[Any] = type_sequence_label_size lowerCamelCase__ : Optional[Any] = initializer_range lowerCamelCase__ : int = num_labels lowerCamelCase__ : List[Any] = num_choices lowerCamelCase__ : int = scope lowerCamelCase__ : Tuple = projection_dim def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : Dict = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py lowerCamelCase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase__ : Tuple = None if self.use_token_type_ids: lowerCamelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : int = None lowerCamelCase__ : Tuple = None if self.use_labels: lowerCamelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase__ : Union[str, Any] = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) lowerCamelCase__ : Tuple = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: List[Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: str , UpperCamelCase__: int , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[Any] = TFDPRContextEncoder(config=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : int = model(UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : str = model(UpperCamelCase__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowerCamelCase_ ( self: int , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[str] , UpperCamelCase__: List[str] , UpperCamelCase__: List[Any] , UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: Any ): lowerCamelCase__ : int = TFDPRQuestionEncoder(config=UpperCamelCase__ ) lowerCamelCase__ : Any = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) lowerCamelCase__ : Tuple = model(UpperCamelCase__ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: str , UpperCamelCase__: List[str] , UpperCamelCase__: Any , UpperCamelCase__: int , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): lowerCamelCase__ : str = TFDPRReader(config=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : Optional[Any] = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : str = config_and_inputs lowerCamelCase__ : Optional[int] = {"""input_ids""": input_ids} return config, inputs_dict @require_tf class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) a = {"""feature-extraction""": TFDPRQuestionEncoder} if is_tf_available() else {} a = False a = False a = False a = False a = False def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = TFDPRModelTester(self ) lowerCamelCase__ : int = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Dict ): self.config_tester.run_common_tests() def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*UpperCamelCase__ ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: Optional[int] ): for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : str = TFDPRContextEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Optional[int] = TFDPRContextEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Tuple = TFDPRQuestionEncoder.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : int = TFDPRReader.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) @require_tf class _lowercase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : List[str] = TFDPRQuestionEncoder.from_pretrained("""facebook/dpr-question_encoder-single-nq-base""" ) lowerCamelCase__ : Optional[Any] = tf.constant( [[101, 7_592, 1_010, 2_003, 2_026, 3_899, 10_140, 1_029, 102]] ) # [CLS] hello, is my dog cute? [SEP] lowerCamelCase__ : Optional[Any] = model(UpperCamelCase__ )[0] # embedding shape = (1, 768) # compare the actual values for a slice. lowerCamelCase__ : Union[str, Any] = tf.constant( [ [ 0.03_236_253, 0.12_753_335, 0.16_818_509, 0.00_279_786, 0.3_896_933, 0.24_264_945, 0.2_178_971, -0.02_335_227, -0.08_481_959, -0.14_324_117, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1e-4 ) )
41
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _A : Tuple =[] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', F'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', F'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qpos_proj.weight', F'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kpos_proj.weight', F'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.weight', F'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', F'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', F'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kpos_proj.weight', F'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.weight', F'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', F'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', F'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', F'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.sa_qpos_proj.bias', F'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_kpos_proj.bias', F'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.sa_v_proj.bias', F'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', F'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', F'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.ca_kpos_proj.bias', F'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.ca_v_proj.bias', F'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', F'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: lowerCamelCase__ : List[Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Dict: lowerCamelCase__ : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: lowerCamelCase__ : List[str] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) lowerCamelCase__ : Optional[int] = value else: lowerCamelCase__ : Any = value return new_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False ) -> Dict: lowerCamelCase__ : Optional[int] = """""" if is_panoptic: lowerCamelCase__ : Dict = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCamelCase__ : List[Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) lowerCamelCase__ : Union[str, Any] = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[:256, :] lowerCamelCase__ : Any = in_proj_bias[:256] lowerCamelCase__ : str = in_proj_weight[256:512, :] lowerCamelCase__ : Optional[int] = in_proj_bias[256:512] lowerCamelCase__ : Dict = in_proj_weight[-256:, :] lowerCamelCase__ : str = in_proj_bias[-256:] def SCREAMING_SNAKE_CASE_ () -> List[Any]: lowerCamelCase__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: lowerCamelCase__ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: lowerCamelCase__ : Any = """resnet101""" if "dc5" in model_name: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : int = """panoptic""" in model_name if is_panoptic: lowerCamelCase__ : List[str] = 250 else: lowerCamelCase__ : int = 91 lowerCamelCase__ : int = """huggingface/label-files""" lowerCamelCase__ : List[str] = """coco-detection-id2label.json""" lowerCamelCase__ : Optional[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : str = idalabel lowerCamelCase__ : List[str] = {v: k for k, v in idalabel.items()} # load image processor lowerCamelCase__ : Optional[int] = """coco_panoptic""" if is_panoptic else """coco_detection""" lowerCamelCase__ : int = ConditionalDetrImageProcessor(format=UpperCamelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Optional[Any] = encoding["""pixel_values"""] logger.info(f'''Converting model {model_name}...''' ) # load original model from torch hub lowerCamelCase__ : List[Any] = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase , pretrained=UpperCamelCase ).eval() lowerCamelCase__ : Dict = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: lowerCamelCase__ : Optional[Any] = """conditional_detr.""" + src rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : Dict = rename_backbone_keys(UpperCamelCase ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase , is_panoptic=UpperCamelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCamelCase__ : Dict = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCamelCase__ : List[str] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: lowerCamelCase__ : int = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Tuple = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Dict = val # finally, create HuggingFace model and load state dict lowerCamelCase__ : Tuple = ConditionalDetrForSegmentation(UpperCamelCase ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() model.push_to_hub(repo_id=UpperCamelCase , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion lowerCamelCase__ : Optional[Any] = conditional_detr(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCamelCase ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : List[Any] =argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _A : Optional[Any] =parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot import BlenderbotTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _A : List[Any] =logging.get_logger(__name__) _A : Tuple ={ '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _A : List[Any] ={ '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } _A : List[Any] ={'''facebook/blenderbot-3B''': 128} class _lowercase ( _lowercase ): a = VOCAB_FILES_NAMES a = PRETRAINED_VOCAB_FILES_MAP a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a = ["""input_ids""", """attention_mask"""] a = BlenderbotTokenizer def __init__( self: Union[str, Any] , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[str]=None , UpperCamelCase__: int=None , UpperCamelCase__: Dict="replace" , UpperCamelCase__: Any="<s>" , UpperCamelCase__: Dict="</s>" , UpperCamelCase__: Any="</s>" , UpperCamelCase__: Union[str, Any]="<s>" , UpperCamelCase__: Tuple="<unk>" , UpperCamelCase__: Union[str, Any]="<pad>" , UpperCamelCase__: Optional[Any]="<mask>" , UpperCamelCase__: Tuple=False , UpperCamelCase__: str=True , **UpperCamelCase__: Optional[Any] , ): super().__init__( UpperCamelCase__ , UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , errors=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase__ : Optional[int] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCamelCase__ ) != add_prefix_space: lowerCamelCase__ : List[str] = getattr(UpperCamelCase__ , pre_tok_state.pop("""type""" ) ) lowerCamelCase__ : Union[str, Any] = add_prefix_space lowerCamelCase__ : Optional[int] = pre_tok_class(**UpperCamelCase__ ) lowerCamelCase__ : Dict = add_prefix_space lowerCamelCase__ : Tuple = """post_processor""" lowerCamelCase__ : Tuple = getattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__ ) if tokenizer_component_instance: lowerCamelCase__ : str = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : str = tuple(state["""sep"""] ) if "cls" in state: lowerCamelCase__ : Optional[Any] = tuple(state["""cls"""] ) lowerCamelCase__ : Optional[int] = False if state.get("""add_prefix_space""" , UpperCamelCase__ ) != add_prefix_space: lowerCamelCase__ : Tuple = add_prefix_space lowerCamelCase__ : Optional[Any] = True if state.get("""trim_offsets""" , UpperCamelCase__ ) != trim_offsets: lowerCamelCase__ : int = trim_offsets lowerCamelCase__ : int = True if changes_to_apply: lowerCamelCase__ : List[Any] = getattr(UpperCamelCase__ , state.pop("""type""" ) ) lowerCamelCase__ : Any = component_class(**UpperCamelCase__ ) setattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__ ) @property # Copied from transformers.models.roberta.tokenization_roberta_fast.RobertaTokenizerFast.mask_token with Roberta->Blenderbot, RoBERTa->Blenderbot def lowerCamelCase_ ( self: str ): if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""" ) return None return str(self._mask_token ) @mask_token.setter def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else value lowerCamelCase__ : int = value def lowerCamelCase_ ( self: Union[str, Any] , *UpperCamelCase__: Optional[Any] , **UpperCamelCase__: Any ): lowerCamelCase__ : List[str] = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] , *UpperCamelCase__: List[Any] , **UpperCamelCase__: int ): lowerCamelCase__ : List[str] = kwargs.get("""is_split_into_words""" , UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Optional[str] = None ): lowerCamelCase__ : Dict = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: List[int] , UpperCamelCase__: Optional[List[int]] = None ): return token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self: Any , UpperCamelCase__: "Conversation" ): lowerCamelCase__ : List[str] = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text ) else: # Generated responses should contain them already. inputs.append(UpperCamelCase__ ) lowerCamelCase__ : str = """ """.join(UpperCamelCase__ ) lowerCamelCase__ : str = self.encode(UpperCamelCase__ ) if len(UpperCamelCase__ ) > self.model_max_length: lowerCamelCase__ : List[Any] = input_ids[-self.model_max_length :] logger.warning(F'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''' ) return input_ids
41
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
1
'''simple docstring''' import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: Any , UpperCamelCase__: int , UpperCamelCase__: Dict=7 , UpperCamelCase__: Union[str, Any]=3 , UpperCamelCase__: Tuple=30 , UpperCamelCase__: Union[str, Any]=400 , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Tuple=None , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: List[str]=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , UpperCamelCase__: List[Any]=True , UpperCamelCase__: Any=1 / 255 , UpperCamelCase__: int=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowerCamelCase__ : Optional[int] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1_333} lowerCamelCase__ : int = parent lowerCamelCase__ : Optional[int] = batch_size lowerCamelCase__ : int = num_channels lowerCamelCase__ : Optional[Any] = min_resolution lowerCamelCase__ : Union[str, Any] = max_resolution lowerCamelCase__ : str = do_resize lowerCamelCase__ : Dict = size lowerCamelCase__ : List[str] = do_normalize lowerCamelCase__ : Optional[int] = image_mean lowerCamelCase__ : List[Any] = image_std lowerCamelCase__ : Tuple = do_rescale lowerCamelCase__ : Tuple = rescale_factor lowerCamelCase__ : str = do_pad def lowerCamelCase_ ( self: int ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def lowerCamelCase_ ( self: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=False ): if not batched: lowerCamelCase__ : List[Any] = image_inputs[0] if isinstance(UpperCamelCase__ , Image.Image ): lowerCamelCase__ , lowerCamelCase__ : Tuple = image.size else: lowerCamelCase__ , lowerCamelCase__ : List[Any] = image.shape[1], image.shape[2] if w < h: lowerCamelCase__ : Optional[int] = int(self.size["""shortest_edge"""] * h / w ) lowerCamelCase__ : int = self.size["""shortest_edge"""] elif w > h: lowerCamelCase__ : Any = self.size["""shortest_edge"""] lowerCamelCase__ : Dict = int(self.size["""shortest_edge"""] * w / h ) else: lowerCamelCase__ : Union[str, Any] = self.size["""shortest_edge"""] lowerCamelCase__ : Union[str, Any] = self.size["""shortest_edge"""] else: lowerCamelCase__ : Optional[Any] = [] for image in image_inputs: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase__ : Optional[Any] = max(UpperCamelCase__ , key=lambda UpperCamelCase__ : item[0] )[0] lowerCamelCase__ : Optional[int] = max(UpperCamelCase__ , key=lambda UpperCamelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DetaImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = DetaImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Any ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: str ): lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_rescale""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1_333} ) self.assertEqual(image_processor.do_pad , UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[int] ): pass def lowerCamelCase_ ( self: Optional[Any] ): # Initialize image_processing lowerCamelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase__ , lowerCamelCase__ : List[str] = self.image_processor_tester.get_expected_values(UpperCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ , lowerCamelCase__ : int = self.image_processor_tester.get_expected_values(UpperCamelCase__ , batched=UpperCamelCase__ ) lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase_ ( self: List[str] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase__ , lowerCamelCase__ : int = self.image_processor_tester.get_expected_values(UpperCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ : str = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.image_processor_tester.get_expected_values(UpperCamelCase__ , batched=UpperCamelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase_ ( self: Any ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : List[str] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase__ , lowerCamelCase__ : List[str] = self.image_processor_tester.get_expected_values(UpperCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ : List[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values lowerCamelCase__ , lowerCamelCase__ : int = self.image_processor_tester.get_expected_values(UpperCamelCase__ , batched=UpperCamelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCamelCase_ ( self: List[Any] ): # prepare image and target lowerCamelCase__ : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: lowerCamelCase__ : Tuple = json.loads(f.read() ) lowerCamelCase__ : Union[str, Any] = {"""image_id""": 39_769, """annotations""": target} # encode them lowerCamelCase__ : List[Any] = DetaImageProcessor() lowerCamelCase__ : Union[str, Any] = image_processing(images=UpperCamelCase__ , annotations=UpperCamelCase__ , return_tensors="""pt""" ) # verify pixel values lowerCamelCase__ : Union[str, Any] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , UpperCamelCase__ ) lowerCamelCase__ : str = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , UpperCamelCase__ , atol=1e-4 ) ) # verify area lowerCamelCase__ : Optional[int] = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , UpperCamelCase__ ) ) # verify boxes lowerCamelCase__ : Optional[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = torch.tensor([0.5_503, 0.2_765, 0.0_604, 0.2_215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , UpperCamelCase__ , atol=1e-3 ) ) # verify image_id lowerCamelCase__ : str = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , UpperCamelCase__ ) ) # verify is_crowd lowerCamelCase__ : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , UpperCamelCase__ ) ) # verify class_labels lowerCamelCase__ : str = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , UpperCamelCase__ ) ) # verify orig_size lowerCamelCase__ : Optional[Any] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , UpperCamelCase__ ) ) # verify size lowerCamelCase__ : List[str] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , UpperCamelCase__ ) ) @slow def lowerCamelCase_ ( self: Optional[int] ): # prepare image, target and masks_path lowerCamelCase__ : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: lowerCamelCase__ : List[Any] = json.loads(f.read() ) lowerCamelCase__ : Dict = {"""file_name""": """000000039769.png""", """image_id""": 39_769, """segments_info""": target} lowerCamelCase__ : Optional[int] = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them lowerCamelCase__ : str = DetaImageProcessor(format="""coco_panoptic""" ) lowerCamelCase__ : Union[str, Any] = image_processing(images=UpperCamelCase__ , annotations=UpperCamelCase__ , masks_path=UpperCamelCase__ , return_tensors="""pt""" ) # verify pixel values lowerCamelCase__ : List[str] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , UpperCamelCase__ ) lowerCamelCase__ : int = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , UpperCamelCase__ , atol=1e-4 ) ) # verify area lowerCamelCase__ : Any = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , UpperCamelCase__ ) ) # verify boxes lowerCamelCase__ : Optional[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = torch.tensor([0.2_625, 0.5_437, 0.4_688, 0.8_625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , UpperCamelCase__ , atol=1e-3 ) ) # verify image_id lowerCamelCase__ : Tuple = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , UpperCamelCase__ ) ) # verify is_crowd lowerCamelCase__ : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , UpperCamelCase__ ) ) # verify class_labels lowerCamelCase__ : Tuple = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , UpperCamelCase__ ) ) # verify masks lowerCamelCase__ : List[str] = 822_873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , UpperCamelCase__ ) # verify orig_size lowerCamelCase__ : Any = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , UpperCamelCase__ ) ) # verify size lowerCamelCase__ : Union[str, Any] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , UpperCamelCase__ ) )
41
'''simple docstring''' class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : dict[str, TrieNode] = {} # Mapping from char to TrieNode lowerCamelCase__ : List[str] = False def lowerCamelCase_ ( self: str , UpperCamelCase__: list[str] ): for word in words: self.insert(UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: str ): lowerCamelCase__ : List[Any] = self for char in word: if char not in curr.nodes: lowerCamelCase__ : Tuple = TrieNode() lowerCamelCase__ : List[Any] = curr.nodes[char] lowerCamelCase__ : Any = True def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: str ): lowerCamelCase__ : Union[str, Any] = self for char in word: if char not in curr.nodes: return False lowerCamelCase__ : Any = curr.nodes[char] return curr.is_leaf def lowerCamelCase_ ( self: str , UpperCamelCase__: str ): def _delete(UpperCamelCase__: TrieNode , UpperCamelCase__: str , UpperCamelCase__: int ) -> bool: if index == len(UpperCamelCase__ ): # If word does not exist if not curr.is_leaf: return False lowerCamelCase__ : str = False return len(curr.nodes ) == 0 lowerCamelCase__ : List[str] = word[index] lowerCamelCase__ : Dict = curr.nodes.get(UpperCamelCase__ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCamelCase__ : List[Any] = _delete(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCamelCase__ , 0 ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: if node.is_leaf: print(UpperCamelCase , end=""" """ ) for key, value in node.nodes.items(): print_words(UpperCamelCase , word + key ) def SCREAMING_SNAKE_CASE_ () -> bool: lowerCamelCase__ : str = """banana bananas bandana band apple all beast""".split() lowerCamelCase__ : Union[str, Any] = TrieNode() root.insert_many(UpperCamelCase ) # print_words(root, "") assert all(root.find(UpperCamelCase ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> None: print(str(UpperCamelCase ) , """works!""" if passes else """doesn't work :(""" ) def SCREAMING_SNAKE_CASE_ () -> None: assert test_trie() def SCREAMING_SNAKE_CASE_ () -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
41
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore _A : Optional[Any] =''' Human: <<task>> Assistant: ''' _A : List[str] ='''huggingface-tools/default-prompts''' _A : int ={'''chat''': '''chat_prompt_template.txt''', '''run''': '''run_prompt_template.txt'''} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: if prompt_or_repo_id is None: lowerCamelCase__ : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("""\\s""" , UpperCamelCase ) is not None: return prompt_or_repo_id lowerCamelCase__ : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="""dataset""" , user_agent={"""agent""": agent_name} ) with open(UpperCamelCase , """r""" , encoding="""utf-8""" ) as f: return f.read()
41
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _A : List[str] =logging.get_logger(__name__) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False ) -> Union[str, Any]: lowerCamelCase__ : str = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''transformer.blocks.{i}.norm1.weight''', f'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm1.bias''', f'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.weight''', f'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (f'''transformer.blocks.{i}.attn.proj.bias''', f'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.weight''', f'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.norm2.bias''', f'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (f'''transformer.blocks.{i}.mlp.fc1.weight''', f'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc1.bias''', f'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.weight''', f'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''transformer.blocks.{i}.mlp.fc2.bias''', f'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[Any]: for i in range(config.num_hidden_layers ): lowerCamelCase__ : str = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : List[str] = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.weight''' ) lowerCamelCase__ : str = state_dict.pop(f'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[: config.hidden_size] lowerCamelCase__ : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[str] = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : Dict = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: lowerCamelCase__ : int = dct.pop(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = val @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False lowerCamelCase__ : int = False if "vqa" in checkpoint_url: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Any = 3129 lowerCamelCase__ : Tuple = """huggingface/label-files""" lowerCamelCase__ : List[str] = """vqa2-id2label.json""" lowerCamelCase__ : str = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase__ : Any = {int(UpperCamelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : int = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Tuple = ViltForQuestionAnswering(UpperCamelCase ) elif "nlvr" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : Any = {0: """False""", 1: """True"""} lowerCamelCase__ : int = {v: k for k, v in config.idalabel.items()} lowerCamelCase__ : Any = 3 lowerCamelCase__ : List[str] = ViltForImagesAndTextClassification(UpperCamelCase ) elif "irtr" in checkpoint_url: lowerCamelCase__ : List[str] = True lowerCamelCase__ : Optional[int] = ViltForImageAndTextRetrieval(UpperCamelCase ) elif "mlm_itm" in checkpoint_url: lowerCamelCase__ : Optional[Any] = True lowerCamelCase__ : Optional[Any] = ViltForMaskedLM(UpperCamelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys lowerCamelCase__ : Dict = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" )["""state_dict"""] lowerCamelCase__ : List[Any] = create_rename_keys(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) for src, dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) read_in_q_k_v(UpperCamelCase , UpperCamelCase ) if mlm_model or irtr_model: lowerCamelCase__ : List[str] = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(UpperCamelCase ) # Define processor lowerCamelCase__ : Optional[int] = ViltImageProcessor(size=384 ) lowerCamelCase__ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) lowerCamelCase__ : Union[str, Any] = ViltProcessor(UpperCamelCase , UpperCamelCase ) # Forward pass on example inputs (image + text) if nlvr_model: lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : int = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Dict = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) lowerCamelCase__ : Optional[int] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Dict = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : List[str] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowerCamelCase__ : str = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=UpperCamelCase ).raw ) if mlm_model: lowerCamelCase__ : str = """a bunch of [MASK] laying on a [MASK].""" else: lowerCamelCase__ : Optional[int] = """How many cats are there?""" lowerCamelCase__ : List[str] = processor(UpperCamelCase , UpperCamelCase , return_tensors="""pt""" ) lowerCamelCase__ : Union[str, Any] = model(**UpperCamelCase ) # Verify outputs if mlm_model: lowerCamelCase__ : Tuple = torch.Size([1, 11, 30522] ) lowerCamelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify masked token prediction equals "cats" lowerCamelCase__ : int = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: lowerCamelCase__ : str = torch.Size([1, 3129] ) lowerCamelCase__ : Any = torch.tensor([-15.9495, -18.1472, -10.3041] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , UpperCamelCase , atol=1E-4 ) # verify vqa prediction equals "2" lowerCamelCase__ : Tuple = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCamelCase__ : str = torch.Size([1, 2] ) lowerCamelCase__ : Optional[Any] = torch.tensor([-2.8721, 2.1291] ) assert torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) assert outputs.logits.shape == expected_shape Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _A : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _A : Tuple =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
41
1
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer _A : Union[str, Any] =['''bert-base-uncased''', '''bert-base-cased'''] _A : int ='''hf-internal-testing/tiny-bert-tf-only''' if is_tf_available(): class _lowercase ( tf.keras.Model ): def __init__( self: Optional[int] , UpperCamelCase__: List[Any] ): super().__init__() lowerCamelCase__ : Optional[int] = tokenizer lowerCamelCase__ : Dict = AutoConfig.from_pretrained(UpperCamelCase__ ) lowerCamelCase__ : Tuple = TFAutoModel.from_config(UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Optional[Any] ): lowerCamelCase__ : Any = self.tokenizer(UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = self.bert(**UpperCamelCase__ ) return out["pooler_output"] @require_tf @require_tensorflow_text class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Any ): super().setUp() lowerCamelCase__ : str = [ BertTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false lowerCamelCase__ : int = [TFBertTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(UpperCamelCase__ , use_fast_bert_tokenizer=UpperCamelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) lowerCamelCase__ : Tuple = [ """This is a straightforward English test sentence.""", """This one has some weird characters\rto\nsee\r\nif those\u00E9break things.""", """Now we're going to add some Chinese: 一 二 三 一二三""", """And some much more rare Chinese: 齉 堃 齉堃""", """Je vais aussi écrire en français pour tester les accents""", """Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ""", ] lowerCamelCase__ : Optional[int] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def lowerCamelCase_ ( self: Optional[Any] ): for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): lowerCamelCase__ : List[Any] = tokenizer(UpperCamelCase__ , return_tensors="""tf""" , padding="""longest""" ) lowerCamelCase__ : List[str] = tf_tokenizer(UpperCamelCase__ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def lowerCamelCase_ ( self: List[Any] ): for tf_tokenizer in self.tf_tokenizers: lowerCamelCase__ : int = tf_tokenizer(self.paired_sentences ) lowerCamelCase__ : int = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def lowerCamelCase_ ( self: Any ): for tf_tokenizer in self.tf_tokenizers: lowerCamelCase__ : List[Any] = tf.function(UpperCamelCase__ ) for test_inputs in (self.test_sentences, self.paired_sentences): lowerCamelCase__ : Optional[int] = tf.constant(UpperCamelCase__ ) lowerCamelCase__ : List[str] = compiled_tokenizer(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = tf_tokenizer(UpperCamelCase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def lowerCamelCase_ ( self: Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: lowerCamelCase__ : Dict = ModelToSave(tokenizer=UpperCamelCase__ ) lowerCamelCase__ : str = tf.convert_to_tensor(self.test_sentences ) lowerCamelCase__ : Any = model(UpperCamelCase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: lowerCamelCase__ : List[Any] = Path(UpperCamelCase__ ) / """saved.model""" model.save(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = tf.keras.models.load_model(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = loaded_model(UpperCamelCase__ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1e-5 )
41
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _A : List[Any] =logging.get_logger(__name__) _A : Dict =['''model.decoder.embed_positions.weights'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> str: if "emb" in name: lowerCamelCase__ : Dict = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: lowerCamelCase__ : List[str] = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: lowerCamelCase__ : List[str] = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: lowerCamelCase__ : Optional[int] = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: lowerCamelCase__ : Union[str, Any] = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: lowerCamelCase__ : Dict = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: lowerCamelCase__ : Optional[Any] = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: lowerCamelCase__ : Dict = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: lowerCamelCase__ : Optional[Any] = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: lowerCamelCase__ : Optional[Any] = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: lowerCamelCase__ : int = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Tuple[Dict, Dict]: lowerCamelCase__ : int = list(state_dict.keys() ) lowerCamelCase__ : Tuple = {} for key in keys: lowerCamelCase__ : Any = state_dict.pop(UpperCamelCase ) lowerCamelCase__ : Union[str, Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj lowerCamelCase__ : Union[str, Any] = val[:hidden_size, :] lowerCamelCase__ : Any = val[hidden_size : 2 * hidden_size, :] lowerCamelCase__ : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: lowerCamelCase__ : str = val else: lowerCamelCase__ : Union[str, Any] = val return state_dict, enc_dec_proj_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values lowerCamelCase__ : int = 1024 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 16 elif checkpoint == "medium": lowerCamelCase__ : Any = 1536 lowerCamelCase__ : Union[str, Any] = 48 lowerCamelCase__ : Optional[int] = 24 elif checkpoint == "large": lowerCamelCase__ : Optional[Any] = 2048 lowerCamelCase__ : Dict = 48 lowerCamelCase__ : List[Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) lowerCamelCase__ : Any = MusicgenDecoderConfig( hidden_size=UpperCamelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase , num_attention_heads=UpperCamelCase , ) return config @torch.no_grad() def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase="cpu" ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = MusicGen.get_pretrained(UpperCamelCase , device=UpperCamelCase ) lowerCamelCase__ : List[Any] = decoder_config_from_checkpoint(UpperCamelCase ) lowerCamelCase__ : Any = fairseq_model.lm.state_dict() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = rename_state_dict( UpperCamelCase , hidden_size=decoder_config.hidden_size ) lowerCamelCase__ : str = TaEncoderModel.from_pretrained("""t5-base""" ) lowerCamelCase__ : Tuple = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) lowerCamelCase__ : Optional[int] = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection lowerCamelCase__ , lowerCamelCase__ : List[str] = decoder.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model lowerCamelCase__ : Optional[Any] = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase , audio_encoder=UpperCamelCase , decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass lowerCamelCase__ : Dict = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) lowerCamelCase__ : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor lowerCamelCase__ : str = AutoTokenizer.from_pretrained("""t5-base""" ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) lowerCamelCase__ : Optional[int] = MusicgenProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids lowerCamelCase__ : Union[str, Any] = 2048 lowerCamelCase__ : List[str] = 2048 # set other default generation config params lowerCamelCase__ : Optional[Any] = int(30 * audio_encoder.config.frame_rate ) lowerCamelCase__ : Union[str, Any] = True lowerCamelCase__ : List[Any] = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _A : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _A : List[str] =parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
41
1
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class _lowercase : def __init__( self: List[str] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any]=13 , UpperCamelCase__: List[str]=7 , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: Dict=True , UpperCamelCase__: Tuple=False , UpperCamelCase__: Optional[Any]=False , UpperCamelCase__: List[Any]=19 , UpperCamelCase__: str=32 , UpperCamelCase__: Tuple=5 , UpperCamelCase__: Any=4 , UpperCamelCase__: Dict=37 , UpperCamelCase__: Dict="gelu" , UpperCamelCase__: Any=0.1 , UpperCamelCase__: str=0.1 , UpperCamelCase__: Optional[int]=512 , UpperCamelCase__: str=16 , UpperCamelCase__: List[str]=2 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: str=3 , UpperCamelCase__: List[str]=4 , UpperCamelCase__: Union[str, Any]=None , ): lowerCamelCase__ : Any = parent lowerCamelCase__ : str = batch_size lowerCamelCase__ : Optional[Any] = seq_length lowerCamelCase__ : List[Any] = is_training lowerCamelCase__ : Optional[int] = use_input_mask lowerCamelCase__ : List[str] = use_token_type_ids lowerCamelCase__ : List[Any] = use_labels lowerCamelCase__ : List[str] = vocab_size lowerCamelCase__ : str = hidden_size lowerCamelCase__ : str = num_hidden_layers lowerCamelCase__ : Optional[int] = num_attention_heads lowerCamelCase__ : Any = intermediate_size lowerCamelCase__ : Any = hidden_act lowerCamelCase__ : Optional[Any] = hidden_dropout_prob lowerCamelCase__ : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase__ : Tuple = max_position_embeddings lowerCamelCase__ : int = type_vocab_size lowerCamelCase__ : Dict = type_sequence_label_size lowerCamelCase__ : int = initializer_range lowerCamelCase__ : List[Any] = num_labels lowerCamelCase__ : List[Any] = num_choices lowerCamelCase__ : Any = scope def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : int = None if self.use_input_mask: lowerCamelCase__ : int = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase__ : List[str] = None lowerCamelCase__ : List[str] = None lowerCamelCase__ : List[Any] = None if self.use_labels: lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase__ : Any = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase__ : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = EsmConfig( vocab_size=33 , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , is_folding_model=UpperCamelCase__ , esmfold_config={"""trunk""": {"""num_blocks""": 2}, """fp16_esm""": False} , ) return config def lowerCamelCase_ ( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: int , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: Any ): lowerCamelCase__ : Dict = EsmForProteinFolding(config=UpperCamelCase__ ).float() model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ ) lowerCamelCase__ : Dict = model(UpperCamelCase__ ) self.parent.assertEqual(result.positions.shape , (8, self.batch_size, self.seq_length, 14, 3) ) self.parent.assertEqual(result.angles.shape , (8, self.batch_size, self.seq_length, 7, 2) ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : int = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : int = config_and_inputs lowerCamelCase__ : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = False a = (EsmForProteinFolding,) if is_torch_available() else () a = () a = {} if is_torch_available() else {} a = False def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = EsmFoldModelTester(self ) lowerCamelCase__ : Optional[Any] = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def lowerCamelCase_ ( self: Tuple ): self.config_tester.run_common_tests() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) @unittest.skip("""Does not support attention outputs""" ) def lowerCamelCase_ ( self: List[Any] ): pass @unittest.skip def lowerCamelCase_ ( self: int ): pass @unittest.skip("""Esm does not support embedding resizing""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip("""Esm does not support embedding resizing""" ) def lowerCamelCase_ ( self: str ): pass @unittest.skip("""ESMFold does not support passing input embeds!""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip("""ESMFold does not support head pruning.""" ) def lowerCamelCase_ ( self: int ): pass @unittest.skip("""ESMFold does not support head pruning.""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip("""ESMFold does not support head pruning.""" ) def lowerCamelCase_ ( self: List[str] ): pass @unittest.skip("""ESMFold does not support head pruning.""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip("""ESMFold does not support head pruning.""" ) def lowerCamelCase_ ( self: Tuple ): pass @unittest.skip("""ESMFold does not output hidden states in the normal way.""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip("""ESMfold does not output hidden states in the normal way.""" ) def lowerCamelCase_ ( self: Union[str, Any] ): pass @unittest.skip("""ESMFold only has one output format.""" ) def lowerCamelCase_ ( self: Dict ): pass @unittest.skip("""This test doesn't work for ESMFold and doesn't test core functionality""" ) def lowerCamelCase_ ( self: Union[str, Any] ): pass @unittest.skip("""ESMFold does not support input chunking.""" ) def lowerCamelCase_ ( self: Any ): pass @unittest.skip("""ESMFold doesn't respect you and it certainly doesn't respect your initialization arguments.""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def lowerCamelCase_ ( self: Any ): pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def lowerCamelCase_ ( self: str ): pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def lowerCamelCase_ ( self: List[Any] ): pass @unittest.skip("""ESMFold doesn't support data parallel.""" ) def lowerCamelCase_ ( self: Tuple ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCamelCase_ ( self: str ): pass @require_torch class _lowercase ( _lowercase ): @slow def lowerCamelCase_ ( self: str ): lowerCamelCase__ : List[Any] = EsmForProteinFolding.from_pretrained("""facebook/esmfold_v1""" ).float() model.eval() lowerCamelCase__ : Optional[int] = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) lowerCamelCase__ : Any = model(UpperCamelCase__ )["""positions"""] lowerCamelCase__ : Dict = torch.tensor([2.5_828, 0.7_993, -10.9_334] , dtype=torch.floataa ) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0] , UpperCamelCase__ , atol=1e-4 ) )
41
'''simple docstring''' _A : Union[str, Any] =range(2, 20 + 1) _A : List[str] =[10**k for k in range(ks[-1] + 1)] _A : dict[int, dict[int, list[list[int]]]] ={} def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: lowerCamelCase__ : List[str] = sum(a_i[j] for j in range(UpperCamelCase , len(UpperCamelCase ) ) ) lowerCamelCase__ : int = sum(a_i[j] * base[j] for j in range(min(len(UpperCamelCase ) , UpperCamelCase ) ) ) lowerCamelCase__ , lowerCamelCase__ : Dict = 0, 0 lowerCamelCase__ : List[str] = n - i lowerCamelCase__ : Optional[Any] = memo.get(UpperCamelCase ) if sub_memo is not None: lowerCamelCase__ : str = sub_memo.get(UpperCamelCase ) if jumps is not None and len(UpperCamelCase ) > 0: # find and make the largest jump without going over lowerCamelCase__ : Optional[Any] = -1 for _k in range(len(UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowerCamelCase__ : Dict = _k break if max_jump >= 0: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = jumps[max_jump] # since the difference between jumps is cached, add c lowerCamelCase__ : Dict = diff + c for j in range(min(UpperCamelCase , len(UpperCamelCase ) ) ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = divmod(UpperCamelCase , 10 ) if new_c > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: lowerCamelCase__ : Any = [] else: lowerCamelCase__ : str = {c: []} lowerCamelCase__ : Tuple = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowerCamelCase__ , lowerCamelCase__ : Dict = next_term(UpperCamelCase , k - 1 , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowerCamelCase__ , lowerCamelCase__ : Optional[int] = compute(UpperCamelCase , UpperCamelCase , i + dn , UpperCamelCase ) diff += _diff dn += terms_jumped lowerCamelCase__ : str = sub_memo[c] # keep jumps sorted by # of terms skipped lowerCamelCase__ : List[Any] = 0 while j < len(UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCamelCase , (diff, dn, k) ) return (diff, dn) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: if i >= n: return 0, i if k > len(UpperCamelCase ): a_i.extend([0 for _ in range(k - len(UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowerCamelCase__ : Optional[Any] = i lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = 0, 0, 0 for j in range(len(UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowerCamelCase__ : Optional[int] = ds_c + ds_b diff += addend lowerCamelCase__ : int = 0 for j in range(UpperCamelCase ): lowerCamelCase__ : str = a_i[j] + addend lowerCamelCase__ , lowerCamelCase__ : int = divmod(UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return diff, i - start_i def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: for j in range(UpperCamelCase , len(UpperCamelCase ) ): lowerCamelCase__ : List[Any] = digits[j] + addend if s >= 10: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = divmod(UpperCamelCase , 10 ) lowerCamelCase__ : Any = addend // 10 + quotient else: lowerCamelCase__ : Any = s lowerCamelCase__ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: lowerCamelCase__ , lowerCamelCase__ : Any = divmod(UpperCamelCase , 10 ) digits.append(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 10**15 ) -> int: lowerCamelCase__ : Any = [1] lowerCamelCase__ : List[str] = 1 lowerCamelCase__ : Tuple = 0 while True: lowerCamelCase__ , lowerCamelCase__ : Any = next_term(UpperCamelCase , 20 , i + dn , UpperCamelCase ) dn += terms_jumped if dn == n - i: break lowerCamelCase__ : Union[str, Any] = 0 for j in range(len(UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'{solution() = }')
41
1
'''simple docstring''' class _lowercase : def __init__( self: Union[str, Any] , UpperCamelCase__: Dict ): lowerCamelCase__ : Union[str, Any] = val lowerCamelCase__ : List[str] = None lowerCamelCase__ : List[Any] = None def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Union[str, Any] ): if self.val: if val < self.val: if self.left is None: lowerCamelCase__ : List[str] = Node(UpperCamelCase__ ) else: self.left.insert(UpperCamelCase__ ) elif val > self.val: if self.right is None: lowerCamelCase__ : Dict = Node(UpperCamelCase__ ) else: self.right.insert(UpperCamelCase__ ) else: lowerCamelCase__ : List[str] = val def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Any: # Recursive traversal if root: inorder(root.left , UpperCamelCase ) res.append(root.val ) inorder(root.right , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: # Build BST if len(UpperCamelCase ) == 0: return arr lowerCamelCase__ : str = Node(arr[0] ) for i in range(1 , len(UpperCamelCase ) ): root.insert(arr[i] ) # Traverse BST in order. lowerCamelCase__ : Union[str, Any] = [] inorder(UpperCamelCase , UpperCamelCase ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
41
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: return abs(UpperCamelCase ) if a == 0 else greatest_common_divisor(b % a , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. lowerCamelCase__ , lowerCamelCase__ : Tuple = y, x % y return abs(UpperCamelCase ) def SCREAMING_SNAKE_CASE_ () -> Tuple: try: lowerCamelCase__ : Dict = input("""Enter two integers separated by comma (,): """ ).split(""",""" ) lowerCamelCase__ : Any = int(nums[0] ) lowerCamelCase__ : Optional[Any] = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(UpperCamelCase , UpperCamelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(UpperCamelCase , UpperCamelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("""Wrong input""" ) if __name__ == "__main__": main()
41
1
'''simple docstring''' import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _lowercase ( _lowercase , _lowercase ): @register_to_config def __init__( self: Optional[Any] , *, UpperCamelCase__: int = 4 , UpperCamelCase__: int = 768 , UpperCamelCase__: int , UpperCamelCase__: int , ): super().__init__() lowerCamelCase__ : Dict = nn.Parameter(torch.zeros(UpperCamelCase__ ) ) # parameters for additional clip time embeddings lowerCamelCase__ : int = nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : Tuple = nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) # parameters for encoder hidden states lowerCamelCase__ : Optional[int] = clip_extra_context_tokens lowerCamelCase__ : Optional[int] = nn.Linear( UpperCamelCase__ , self.clip_extra_context_tokens * cross_attention_dim ) lowerCamelCase__ : List[Any] = nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase__ : List[Any] = nn.LayerNorm(UpperCamelCase__ ) def lowerCamelCase_ ( self: int , *, UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str , UpperCamelCase__: int , UpperCamelCase__: Union[str, Any] ): if do_classifier_free_guidance: # Add the classifier free guidance embeddings to the image embeddings lowerCamelCase__ : Dict = image_embeddings.shape[0] lowerCamelCase__ : Optional[int] = self.learned_classifier_free_guidance_embeddings.unsqueeze(0 ) lowerCamelCase__ : List[str] = classifier_free_guidance_embeddings.expand( UpperCamelCase__ , -1 ) lowerCamelCase__ : str = torch.cat([classifier_free_guidance_embeddings, image_embeddings] , dim=0 ) # The image embeddings batch size and the text embeddings batch size are equal assert image_embeddings.shape[0] == prompt_embeds.shape[0] lowerCamelCase__ : Dict = prompt_embeds.shape[0] # "Specifically, we modify the architecture described in Nichol et al. (2021) by projecting and # adding CLIP embeddings to the existing timestep embedding, ... lowerCamelCase__ : List[Any] = self.embedding_proj(UpperCamelCase__ ) lowerCamelCase__ : List[str] = self.clip_image_embeddings_project_to_time_embeddings(UpperCamelCase__ ) lowerCamelCase__ : Dict = time_projected_image_embeddings + time_projected_prompt_embeds # ... and by projecting CLIP embeddings into four # extra tokens of context that are concatenated to the sequence of outputs from the GLIDE text encoder" lowerCamelCase__ : List[str] = self.clip_extra_context_tokens_proj(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = clip_extra_context_tokens.reshape(UpperCamelCase__ , -1 , self.clip_extra_context_tokens ) lowerCamelCase__ : int = clip_extra_context_tokens.permute(0 , 2 , 1 ) lowerCamelCase__ : Optional[int] = self.encoder_hidden_states_proj(UpperCamelCase__ ) lowerCamelCase__ : str = self.text_encoder_hidden_states_norm(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = torch.cat([clip_extra_context_tokens, text_encoder_hidden_states] , dim=1 ) return text_encoder_hidden_states, additive_clip_time_embeddings
41
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self: str , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any=7 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[str]=18 , UpperCamelCase__: Union[str, Any]=30 , UpperCamelCase__: List[str]=400 , UpperCamelCase__: Any=True , UpperCamelCase__: Union[str, Any]=None , UpperCamelCase__: List[Any]=True , UpperCamelCase__: List[Any]=False , UpperCamelCase__: Tuple=True , UpperCamelCase__: Tuple=True , UpperCamelCase__: Any=[0.5, 0.5, 0.5] , UpperCamelCase__: Optional[Any]=[0.5, 0.5, 0.5] , ): lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Union[str, Any] = image_size lowerCamelCase__ : Optional[int] = min_resolution lowerCamelCase__ : Optional[Any] = max_resolution lowerCamelCase__ : Union[str, Any] = do_resize lowerCamelCase__ : Optional[Any] = size if size is not None else {"""height""": 18, """width""": 20} lowerCamelCase__ : Dict = do_thumbnail lowerCamelCase__ : Optional[int] = do_align_axis lowerCamelCase__ : Any = do_pad lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Union[str, Any] = image_mean lowerCamelCase__ : Union[str, Any] = image_std def lowerCamelCase_ ( self: str ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _lowercase ( _lowercase , unittest.TestCase ): a = DonutImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = DonutImageProcessingTester(self ) @property def lowerCamelCase_ ( self: Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_thumbnail""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_align_long_axis""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_pad""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """image_std""" ) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 20} ) lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) # Previous config had dimensions in (width, height) order lowerCamelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {"""height""": 84, """width""": 42} ) def lowerCamelCase_ ( self: List[str] ): pass @is_flaky() def lowerCamelCase_ ( self: Union[str, Any] ): # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , Image.Image ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : List[str] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Optional[int] ): # Initialize image_processing lowerCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , numpify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , np.ndarray ) # Test not batched input lowerCamelCase__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) @is_flaky() def lowerCamelCase_ ( self: Dict ): # Initialize image_processing lowerCamelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase__ , torchify=UpperCamelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
41
1
'''simple docstring''' from typing import Any class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Any ): lowerCamelCase__ : Dict = data lowerCamelCase__ : Optional[int] = None class _lowercase : def __init__( self: Optional[Any] ): lowerCamelCase__ : int = None def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[int] = self.head while temp is not None: print(temp.data , end=""" """ ) lowerCamelCase__ : str = temp.next print() def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Any ): lowerCamelCase__ : Optional[Any] = Node(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.head lowerCamelCase__ : int = new_node def lowerCamelCase_ ( self: int , UpperCamelCase__: str , UpperCamelCase__: Any ): if node_data_a == node_data_a: return else: lowerCamelCase__ : str = self.head while node_a is not None and node_a.data != node_data_a: lowerCamelCase__ : Tuple = node_a.next lowerCamelCase__ : Optional[Any] = self.head while node_a is not None and node_a.data != node_data_a: lowerCamelCase__ : int = node_a.next if node_a is None or node_a is None: return lowerCamelCase__ , lowerCamelCase__ : Dict = node_a.data, node_a.data if __name__ == "__main__": _A : Dict =LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _A : Dict ={'''tokenization_herbert''': ['''HerbertTokenizer''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''HerbertTokenizerFast'''] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _A : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _lowercase : def __init__( self: List[Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Dict=2 , UpperCamelCase__: Optional[int]=True , UpperCamelCase__: Any=False , UpperCamelCase__: Dict=10 , UpperCamelCase__: List[str]=3 , UpperCamelCase__: int=32 * 4 , UpperCamelCase__: List[str]=32 * 6 , UpperCamelCase__: Optional[int]=4 , UpperCamelCase__: int=32 , ): lowerCamelCase__ : List[str] = parent lowerCamelCase__ : List[Any] = batch_size lowerCamelCase__ : Union[str, Any] = is_training lowerCamelCase__ : List[Any] = use_auxiliary_loss lowerCamelCase__ : List[Any] = num_queries lowerCamelCase__ : List[str] = num_channels lowerCamelCase__ : Optional[Any] = min_size lowerCamelCase__ : List[Any] = max_size lowerCamelCase__ : int = num_labels lowerCamelCase__ : Optional[int] = mask_feature_size def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Any = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCamelCase__ ) lowerCamelCase__ : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCamelCase__ ) > 0.5 ).float() lowerCamelCase__ : Tuple = (torch.rand((self.batch_size, self.num_labels) , device=UpperCamelCase__ ) > 0.5).long() lowerCamelCase__ : Tuple = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowerCamelCase_ ( self: Optional[int] ): return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.prepare_config_and_inputs() lowerCamelCase__ : Tuple = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowerCamelCase_ ( self: Dict , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = output.encoder_hidden_states lowerCamelCase__ : List[Any] = output.pixel_decoder_hidden_states lowerCamelCase__ : str = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCamelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCamelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCamelCase__ ) , config.decoder_config.decoder_layers ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[int]=False ): with torch.no_grad(): lowerCamelCase__ : Optional[int] = MaskFormerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : Dict = model(pixel_values=UpperCamelCase__ , pixel_mask=UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , output_hidden_states=UpperCamelCase__ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: int , UpperCamelCase__: Dict , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any , UpperCamelCase__: List[str] ): lowerCamelCase__ : List[Any] = MaskFormerForInstanceSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() def comm_check_on_output(UpperCamelCase__: List[str] ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): lowerCamelCase__ : Optional[int] = model(pixel_values=UpperCamelCase__ , pixel_mask=UpperCamelCase__ ) lowerCamelCase__ : Any = model(UpperCamelCase__ ) comm_check_on_output(UpperCamelCase__ ) lowerCamelCase__ : Optional[int] = model( pixel_values=UpperCamelCase__ , pixel_mask=UpperCamelCase__ , mask_labels=UpperCamelCase__ , class_labels=UpperCamelCase__ ) comm_check_on_output(UpperCamelCase__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () a = ( {"""feature-extraction""": MaskFormerModel, """image-segmentation""": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Any = MaskFormerModelTester(self ) lowerCamelCase__ : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): self.config_tester.run_common_tests() def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCamelCase__ , **UpperCamelCase__ , output_hidden_states=UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCamelCase__ ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowerCamelCase_ ( self: str ): pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowerCamelCase_ ( self: str ): pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowerCamelCase_ ( self: Union[str, Any] ): pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowerCamelCase_ ( self: Any ): pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCamelCase_ ( self: List[Any] ): pass def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = model_class(UpperCamelCase__ ) lowerCamelCase__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : int = [*signature.parameters.keys()] lowerCamelCase__ : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in ["facebook/maskformer-swin-small-coco"]: lowerCamelCase__ : Any = MaskFormerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Union[str, Any] = (self.model_tester.min_size,) * 2 lowerCamelCase__ : Dict = { """pixel_values""": torch.randn((2, 3, *size) , device=UpperCamelCase__ ), """mask_labels""": torch.randn((2, 10, *size) , device=UpperCamelCase__ ), """class_labels""": torch.zeros(2 , 10 , device=UpperCamelCase__ ).long(), } lowerCamelCase__ : List[str] = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ ) self.assertTrue(outputs.loss is not None ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCamelCase__ , **UpperCamelCase__ , output_hidden_states=UpperCamelCase__ ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : str = model_class(UpperCamelCase__ ).to(UpperCamelCase__ ) lowerCamelCase__ : List[str] = model(**UpperCamelCase__ , output_attentions=UpperCamelCase__ ) self.assertTrue(outputs.attentions is not None ) def lowerCamelCase_ ( self: Union[str, Any] ): if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss lowerCamelCase__ : Optional[Any] = self.all_model_classes[1] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() lowerCamelCase__ : Optional[int] = model(UpperCamelCase__ , mask_labels=UpperCamelCase__ , class_labels=UpperCamelCase__ ).loss loss.backward() def lowerCamelCase_ ( self: str ): # only MaskFormerForInstanceSegmentation has the loss lowerCamelCase__ : List[str] = self.all_model_classes[1] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : List[str] = True lowerCamelCase__ : Tuple = True lowerCamelCase__ : int = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() lowerCamelCase__ : Tuple = model(UpperCamelCase__ , mask_labels=UpperCamelCase__ , class_labels=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() lowerCamelCase__ : Optional[int] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't lowerCamelCase__ : int = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() lowerCamelCase__ : str = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCamelCase__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) _A : Optional[int] =1e-4 def SCREAMING_SNAKE_CASE_ () -> Optional[int]: lowerCamelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Dict ): return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Optional[Any] = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(UpperCamelCase__ ) lowerCamelCase__ : str = self.default_image_processor lowerCamelCase__ : Dict = prepare_img() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Any = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCamelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): lowerCamelCase__ : List[str] = model(**UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = torch.tensor( [[-0.0_482, 0.9_228, 0.4_951], [-0.2_547, 0.8_017, 0.8_527], [-0.0_069, 0.3_385, -0.0_089]] ).to(UpperCamelCase__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) lowerCamelCase__ : Optional[Any] = torch.tensor( [[-0.8_422, -0.8_434, -0.9_718], [-1.0_144, -0.5_565, -0.4_195], [-1.0_038, -0.4_484, -0.1_961]] ).to(UpperCamelCase__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) lowerCamelCase__ : List[str] = torch.tensor( [[0.2_852, -0.0_159, 0.9_735], [0.6_254, 0.1_858, 0.8_529], [-0.0_680, -0.4_116, 1.8_413]] ).to(UpperCamelCase__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Union[str, Any] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCamelCase__ ) .eval() ) lowerCamelCase__ : Optional[int] = self.default_image_processor lowerCamelCase__ : List[Any] = prepare_img() lowerCamelCase__ : Tuple = image_processor(UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) lowerCamelCase__ : List[Any] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCamelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): lowerCamelCase__ : Dict = model(**UpperCamelCase__ ) # masks_queries_logits lowerCamelCase__ : List[str] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowerCamelCase__ : int = [ [-1.3_737_124, -1.7_724_937, -1.9_364_233], [-1.5_977_281, -1.9_867_939, -2.1_523_695], [-1.5_795_398, -1.9_269_832, -2.093_942], ] lowerCamelCase__ : int = torch.tensor(UpperCamelCase__ ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) # class_queries_logits lowerCamelCase__ : Any = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowerCamelCase__ : Tuple = torch.tensor( [ [1.6_512e00, -5.2_572e00, -3.3_519e00], [3.6_169e-02, -5.9_025e00, -2.9_313e00], [1.0_766e-04, -7.7_630e00, -5.1_263e00], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) def lowerCamelCase_ ( self: List[Any] ): lowerCamelCase__ : Optional[int] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(UpperCamelCase__ ) .eval() ) lowerCamelCase__ : str = self.default_image_processor lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : List[Any] = image_processor(UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) lowerCamelCase__ : List[str] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCamelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): lowerCamelCase__ : Any = model(**UpperCamelCase__ ) # masks_queries_logits lowerCamelCase__ : int = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowerCamelCase__ : Optional[Any] = [[-0.9_046, -2.6_366, -4.6_062], [-3.4_179, -5.7_890, -8.8_057], [-4.9_179, -7.6_560, -10.7_711]] lowerCamelCase__ : List[Any] = torch.tensor(UpperCamelCase__ ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) # class_queries_logits lowerCamelCase__ : Dict = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowerCamelCase__ : Optional[int] = torch.tensor( [[4.7_188, -3.2_585, -2.8_857], [6.6_871, -2.9_181, -1.2_487], [7.2_449, -2.2_764, -2.1_874]] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase__ , atol=UpperCamelCase__ ) ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Optional[int] = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCamelCase__ ) .eval() ) lowerCamelCase__ : Any = self.default_image_processor lowerCamelCase__ : Optional[Any] = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) lowerCamelCase__ : Optional[int] = inputs["""pixel_values"""].to(UpperCamelCase__ ) lowerCamelCase__ : int = [el.to(UpperCamelCase__ ) for el in inputs["""mask_labels"""]] lowerCamelCase__ : int = [el.to(UpperCamelCase__ ) for el in inputs["""class_labels"""]] with torch.no_grad(): lowerCamelCase__ : List[Any] = model(**UpperCamelCase__ ) self.assertTrue(outputs.loss is not None )
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1