code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """s-JoL/Open-Llama-V1""": """https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Optional[int] ="""open-llama""" def __init__( self , UpperCAmelCase_=10_00_00 , UpperCAmelCase_=40_96 , UpperCAmelCase_=1_10_08 , UpperCAmelCase_=32 , UpperCAmelCase_=32 , UpperCAmelCase_="silu" , UpperCAmelCase_=20_48 , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-6 , UpperCAmelCase_=True , UpperCAmelCase_=0 , UpperCAmelCase_=1 , UpperCAmelCase_=2 , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=None , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = max_position_embeddings lowerCAmelCase = hidden_size lowerCAmelCase = intermediate_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_act lowerCAmelCase = initializer_range lowerCAmelCase = rms_norm_eps lowerCAmelCase = use_cache lowerCAmelCase = kwargs.pop( '''use_memorry_efficient_attention''' , __A ) lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_dropout_prob lowerCAmelCase = use_stable_embedding lowerCAmelCase = shared_input_output_embedding lowerCAmelCase = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , tie_word_embeddings=__A , **__A , ) def __snake_case ( self ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __A ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""" ) lowerCAmelCase = self.rope_scaling.get('''type''' , __A ) lowerCAmelCase = self.rope_scaling.get('''factor''' , __A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(__A , __A ) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
717
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase ( _snake_case = 3 ): if isinstance(_snake_case , _snake_case ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(_snake_case ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) lowerCAmelCase = QuantumRegister(_snake_case , '''qr''' ) lowerCAmelCase = ClassicalRegister(_snake_case , '''cr''' ) lowerCAmelCase = QuantumCircuit(_snake_case , _snake_case ) lowerCAmelCase = number_of_qubits for i in range(_snake_case ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(_snake_case ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , _snake_case , _snake_case ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(_snake_case , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(_snake_case , _snake_case ) # simulate with 10000 shots lowerCAmelCase = Aer.get_backend('''qasm_simulator''' ) lowerCAmelCase = execute(_snake_case , _snake_case , shots=10000 ) return job.result().get_counts(_snake_case ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
33
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __UpperCamelCase ( __snake_case ): '''simple docstring''' __a : Optional[Any] =["""image_processor""", """tokenizer"""] __a : Optional[Any] ="""ChineseCLIPImageProcessor""" __a : Any =("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ ): lowerCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _lowercase , ) lowerCAmelCase = kwargs.pop('''feature_extractor''' ) lowerCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_lowercase , _lowercase ) lowerCAmelCase = self.image_processor def __call__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ ): if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: lowerCAmelCase = self.tokenizer(_lowercase , return_tensors=_lowercase , **_lowercase ) if images is not None: lowerCAmelCase = self.image_processor(_lowercase , return_tensors=_lowercase , **_lowercase ) if text is not None and images is not None: lowerCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowercase ) , tensor_type=_lowercase ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.decode(*_lowercase , **_lowercase ) @property def __snake_case ( self ): lowerCAmelCase = self.tokenizer.model_input_names lowerCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def __snake_case ( self ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _lowercase , ) return self.image_processor_class
718
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Any =1 @register_to_config def __init__( self , UpperCAmelCase_=20_00 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=1E-3 ): lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = torch.linspace(1 , self.config.sampling_eps , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCAmelCase = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCAmelCase = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCAmelCase = std.flatten() while len(std.shape ) < len(score.shape ): lowerCAmelCase = std.unsqueeze(-1 ) lowerCAmelCase = -score / std # compute lowerCAmelCase = -1.0 / len(self.timesteps ) lowerCAmelCase = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCAmelCase = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCAmelCase = beta_t.unsqueeze(-1 ) lowerCAmelCase = -0.5 * beta_t * x lowerCAmelCase = torch.sqrt(UpperCAmelCase_ ) lowerCAmelCase = drift - diffusion**2 * score lowerCAmelCase = x + drift * dt # add noise lowerCAmelCase = randn_tensor(x.shape , layout=x.layout , generator=UpperCAmelCase_ , device=x.device , dtype=x.dtype ) lowerCAmelCase = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ): return self.config.num_train_timesteps
33
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : torch.FloatTensor class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' @register_to_config def __init__( self , UpperCAmelCase_ = 6_55_36 , UpperCAmelCase_ = None , UpperCAmelCase_ = 2 , UpperCAmelCase_ = 2 , UpperCAmelCase_ = 0 , UpperCAmelCase_ = "fourier" , UpperCAmelCase_ = True , UpperCAmelCase_ = False , UpperCAmelCase_ = 0.0 , UpperCAmelCase_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase_ = "UNetMidBlock1D" , UpperCAmelCase_ = None , UpperCAmelCase_ = (32, 32, 64) , UpperCAmelCase_ = None , UpperCAmelCase_ = 8 , UpperCAmelCase_ = 1 , UpperCAmelCase_ = False , ): super().__init__() lowerCAmelCase = sample_size # time if time_embedding_type == "fourier": lowerCAmelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A__ , log=A__ , flip_sin_to_cos=A__ ) lowerCAmelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCAmelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=A__ , downscale_freq_shift=A__ ) lowerCAmelCase = block_out_channels[0] if use_timestep_embedding: lowerCAmelCase = block_out_channels[0] * 4 lowerCAmelCase = TimestepEmbedding( in_channels=A__ , time_embed_dim=A__ , act_fn=A__ , out_dim=block_out_channels[0] , ) lowerCAmelCase = nn.ModuleList([] ) lowerCAmelCase = None lowerCAmelCase = nn.ModuleList([] ) lowerCAmelCase = None # down lowerCAmelCase = in_channels for i, down_block_type in enumerate(A__ ): lowerCAmelCase = output_channel lowerCAmelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCAmelCase = i == len(A__ ) - 1 lowerCAmelCase = get_down_block( A__ , num_layers=A__ , in_channels=A__ , out_channels=A__ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A__ ) # mid lowerCAmelCase = get_mid_block( A__ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A__ , add_downsample=A__ , ) # up lowerCAmelCase = list(reversed(A__ ) ) lowerCAmelCase = reversed_block_out_channels[0] if out_block_type is None: lowerCAmelCase = out_channels else: lowerCAmelCase = block_out_channels[0] for i, up_block_type in enumerate(A__ ): lowerCAmelCase = output_channel lowerCAmelCase = ( reversed_block_out_channels[i + 1] if i < len(A__ ) - 1 else final_upsample_channels ) lowerCAmelCase = i == len(A__ ) - 1 lowerCAmelCase = get_up_block( A__ , num_layers=A__ , in_channels=A__ , out_channels=A__ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A__ ) lowerCAmelCase = output_channel # out lowerCAmelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCAmelCase = get_out_block( out_block_type=A__ , num_groups_out=A__ , embed_dim=block_out_channels[0] , out_channels=A__ , act_fn=A__ , fc_dim=block_out_channels[-1] // 4 , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = True , ): lowerCAmelCase = timestep if not torch.is_tensor(A__ ): lowerCAmelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A__ ) and len(timesteps.shape ) == 0: lowerCAmelCase = timesteps[None].to(sample.device ) lowerCAmelCase = self.time_proj(A__ ) if self.config.use_timestep_embedding: lowerCAmelCase = self.time_mlp(A__ ) else: lowerCAmelCase = timestep_embed[..., None] lowerCAmelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCAmelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCAmelCase = () for downsample_block in self.down_blocks: lowerCAmelCase = downsample_block(hidden_states=A__ , temb=A__ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCAmelCase = self.mid_block(A__ , A__ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCAmelCase = down_block_res_samples[-1:] lowerCAmelCase = down_block_res_samples[:-1] lowerCAmelCase = upsample_block(A__ , res_hidden_states_tuple=A__ , temb=A__ ) # 5. post-process if self.out_block: lowerCAmelCase = self.out_block(A__ , A__ ) if not return_dict: return (sample,) return UNetaDOutput(sample=A__ )
719
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
33
0
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) UpperCAmelCase_ ={ """iou_prediction_head.layers.0""": """iou_prediction_head.proj_in""", """iou_prediction_head.layers.1""": """iou_prediction_head.layers.0""", """iou_prediction_head.layers.2""": """iou_prediction_head.proj_out""", """mask_decoder.output_upscaling.0""": """mask_decoder.upscale_conv1""", """mask_decoder.output_upscaling.1""": """mask_decoder.upscale_layer_norm""", """mask_decoder.output_upscaling.3""": """mask_decoder.upscale_conv2""", """mask_downscaling.0""": """mask_embed.conv1""", """mask_downscaling.1""": """mask_embed.layer_norm1""", """mask_downscaling.3""": """mask_embed.conv2""", """mask_downscaling.4""": """mask_embed.layer_norm2""", """mask_downscaling.6""": """mask_embed.conv3""", """point_embeddings""": """point_embed""", """pe_layer.positional_encoding_gaussian_matrix""": """shared_embedding.positional_embedding""", """image_encoder""": """vision_encoder""", """neck.0""": """neck.conv1""", """neck.1""": """neck.layer_norm1""", """neck.2""": """neck.conv2""", """neck.3""": """neck.layer_norm2""", """patch_embed.proj""": """patch_embed.projection""", """.norm""": """.layer_norm""", """blocks""": """layers""", } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = {} state_dict.pop('''pixel_mean''' , lowerCAmelCase__ ) state_dict.pop('''pixel_std''' , lowerCAmelCase__ ) lowerCAmelCase = R'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase = key.replace(lowerCAmelCase__ , lowerCAmelCase__ ) if re.match(lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase = int(re.match(lowerCAmelCase__ , lowerCAmelCase__ ).group(2 ) ) if layer_nb == 0: lowerCAmelCase = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: lowerCAmelCase = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: lowerCAmelCase = key.replace('''layers.2''' , '''proj_out''' ) lowerCAmelCase = value lowerCAmelCase = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case="ybelkada/segment-anything" ): lowerCAmelCase = hf_hub_download(lowerCAmelCase__ , F"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: lowerCAmelCase = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase = SamConfig( vision_config=lowerCAmelCase__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase = SamConfig( vision_config=lowerCAmelCase__ , ) lowerCAmelCase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowerCAmelCase = replace_keys(lowerCAmelCase__ ) lowerCAmelCase = SamImageProcessor() lowerCAmelCase = SamProcessor(image_processor=lowerCAmelCase__ ) lowerCAmelCase = SamModel(lowerCAmelCase__ ) hf_model.load_state_dict(lowerCAmelCase__ ) lowerCAmelCase = hf_model.to('''cuda''' ) lowerCAmelCase = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' lowerCAmelCase = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ).convert('''RGB''' ) lowerCAmelCase = [[[400, 650]]] lowerCAmelCase = [[1]] lowerCAmelCase = processor(images=np.array(lowerCAmelCase__ ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**lowerCAmelCase__ ) lowerCAmelCase = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 lowerCAmelCase = processor( images=np.array(lowerCAmelCase__ ) , input_points=lowerCAmelCase__ , input_labels=lowerCAmelCase__ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**lowerCAmelCase__ ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 lowerCAmelCase = ((75, 275, 1725, 850),) lowerCAmelCase = processor(images=np.array(lowerCAmelCase__ ) , input_boxes=lowerCAmelCase__ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**lowerCAmelCase__ ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. lowerCAmelCase = [[[400, 650], [800, 650]]] lowerCAmelCase = [[1, 1]] lowerCAmelCase = processor( images=np.array(lowerCAmelCase__ ) , input_points=lowerCAmelCase__ , input_labels=lowerCAmelCase__ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**lowerCAmelCase__ ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() UpperCAmelCase_ =["""sam_vit_b_01ec64""", """sam_vit_h_4b8939""", """sam_vit_l_0b3195"""] parser.add_argument( """--model_name""", default="""sam_vit_h_4b8939""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) parser.add_argument( """--model_hub_id""", default="""ybelkada/segment-anything""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) UpperCAmelCase_ =parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
720
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : str =MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = VideoClassificationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ , top_k=2 ) lowerCAmelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): for example in examples: lowerCAmelCase = video_classifier(UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, ] , ) @require_torch def __snake_case ( self ): lowerCAmelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' lowerCAmelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) lowerCAmelCase = pipeline( '''video-classification''' , model=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = video_classifier(UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __snake_case ( self ): pass
33
0
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Dict =IFInpaintingSuperResolutionPipeline __a : int =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} __a : Union[str, Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) __a : Union[str, Any] =PipelineTesterMixin.required_optional_params - {"latents"} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCamelCase__ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCamelCase__ ) else: lowerCAmelCase = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
721
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any ="""hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __snake_case ( self , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCAmelCase_ ) ) lowerCAmelCase = np.random.RandomState(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # warmup pass to apply optimizations lowerCAmelCase = pipe(**self.get_dummy_inputs() ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @property def __snake_case ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __snake_case ( self ): lowerCAmelCase = ort.SessionOptions() lowerCAmelCase = False return options def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCAmelCase_ , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
33
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
700
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCAmelCase ( _snake_case ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('''/''' ) lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase = torch.load(os.path.join(_snake_case , '''pytorch_model.bin''' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase , lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_snake_case ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F"""bertarized_{os.path.basename(_snake_case )}""" ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(_snake_case , os.path.join(_snake_case , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) UpperCAmelCase_ =parser.parse_args() main(args)
33
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """google/bit-50""": """https://huggingface.co/google/bit-50/resolve/main/config.json""", } class __UpperCamelCase ( _UpperCAmelCase , _UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] ="""bit""" __a : Dict =["""preactivation""", """bottleneck"""] __a : List[Any] =["""SAME""", """VALID"""] def __init__( self , UpperCAmelCase_=3 , UpperCAmelCase_=64 , UpperCAmelCase_=[2_56, 5_12, 10_24, 20_48] , UpperCAmelCase_=[3, 4, 6, 3] , UpperCAmelCase_="preactivation" , UpperCAmelCase_="relu" , UpperCAmelCase_=None , UpperCAmelCase_=32 , UpperCAmelCase_=0.0 , UpperCAmelCase_=False , UpperCAmelCase_=32 , UpperCAmelCase_=1 , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ , ): super().__init__(**lowercase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {",".join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: lowerCAmelCase = global_padding.upper() else: raise ValueError(F"""Padding strategy {global_padding} not supported""" ) lowerCAmelCase = num_channels lowerCAmelCase = embedding_size lowerCAmelCase = hidden_sizes lowerCAmelCase = depths lowerCAmelCase = layer_type lowerCAmelCase = hidden_act lowerCAmelCase = global_padding lowerCAmelCase = num_groups lowerCAmelCase = drop_path_rate lowerCAmelCase = embedding_dynamic_padding lowerCAmelCase = output_stride lowerCAmelCase = width_factor lowerCAmelCase = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowercase__ ) + 1 )] lowerCAmelCase = get_aligned_output_features_output_indices( out_features=lowercase__ , out_indices=lowercase__ , stage_names=self.stage_names )
701
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } UpperCAmelCase_ ={ """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } UpperCAmelCase_ ={ """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = set() lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase = char lowerCAmelCase = set(_snake_case ) return pairs class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =VOCAB_FILES_NAMES __a : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP __a : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , **UpperCAmelCase_ , ): super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = merges_file lowerCAmelCase = {} lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 lowerCAmelCase = 3 self.add_from_file(UpperCAmelCase_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase_ , encoding='''utf-8''' ) as merges_handle: lowerCAmelCase = merges_handle.read().split('''\n''' )[:-1] lowerCAmelCase = [tuple(merge.split()[:-1] ) for merge in merges] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __snake_case ( self ): return len(self.encoder ) def __snake_case ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self , UpperCAmelCase_ ): if token in self.cache: return self.cache[token] lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowerCAmelCase = get_pairs(UpperCAmelCase_ ) if not pairs: return token while True: lowerCAmelCase = min(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : self.bpe_ranks.get(UpperCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase = bigram lowerCAmelCase = [] lowerCAmelCase = 0 while i < len(UpperCAmelCase_ ): try: lowerCAmelCase = word.index(UpperCAmelCase_ , UpperCAmelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase = j if word[i] == first and i < len(UpperCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = new_word if len(UpperCAmelCase_ ) == 1: break else: lowerCAmelCase = get_pairs(UpperCAmelCase_ ) lowerCAmelCase = '''@@ '''.join(UpperCAmelCase_ ) lowerCAmelCase = word[:-4] lowerCAmelCase = word return word def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [] lowerCAmelCase = re.findall(r'''\S+\n?''' , UpperCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __snake_case ( self , UpperCAmelCase_ ): return self.encoder.get(UpperCAmelCase_ , self.encoder.get(self.unk_token ) ) def __snake_case ( self , UpperCAmelCase_ ): return self.decoder.get(UpperCAmelCase_ , self.unk_token ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = ''' '''.join(UpperCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.merges_file , UpperCAmelCase_ ) return out_vocab_file, out_merge_file def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(UpperCAmelCase_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCAmelCase = f.readlines() for lineTmp in lines: lowerCAmelCase = lineTmp.strip() lowerCAmelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowerCAmelCase = line[:idx] lowerCAmelCase = len(self.encoder )
33
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase_ ={"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
702
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ =TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = self lowerCAmelCase = 0 class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # create a new set with x as its member lowerCAmelCase = DisjointSetTreeNode(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): # find the set x belongs to (with path-compression) lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # helper function for union operation if nodea.rank > nodea.rank: lowerCAmelCase = nodea else: lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # merge 2 disjoint sets self.link(self.find_set(UpperCAmelCase_ ) , self.find_set(UpperCAmelCase_ ) ) class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # add a node ONLY if its not present in the graph if node not in self.connections: lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # add an edge with the given weight self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) lowerCAmelCase = weight lowerCAmelCase = weight def __snake_case ( self ): lowerCAmelCase = [] lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCAmelCase_ : x[2] ) # creating the disjoint set lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(UpperCAmelCase_ ) # MST generation lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = edges[index] index += 1 lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) disjoint_set.union(UpperCAmelCase_ , UpperCAmelCase_ ) return graph
33
0
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( __A ): '''simple docstring''' def __init__( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , UpperCAmelCase_ , ) super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ )
703
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations(_snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations_with_dp_array( _snake_case , _snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _snake_case ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =3 UpperCAmelCase_ =5 UpperCAmelCase_ =[1, 2, 5] print(combination_sum_iv(n, array, target))
33
0
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() UpperCAmelCase_ =2 class __UpperCamelCase : '''simple docstring''' def __init__( self , *, # begin keyword-only arguments UpperCAmelCase_="<s>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_=None , ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = bos, unk, pad, eos lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = {} lowerCAmelCase = self.add_symbol(_A ) lowerCAmelCase = self.add_symbol(_A ) lowerCAmelCase = self.add_symbol(_A ) lowerCAmelCase = self.add_symbol(_A ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(_A ) lowerCAmelCase = len(self.symbols ) def __eq__( self , UpperCAmelCase_ ): return self.indices == other.indices def __getitem__( self , UpperCAmelCase_ ): if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): return len(self.symbols ) def __contains__( self , UpperCAmelCase_ ): return sym in self.indices @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = cls() d.add_from_file(_A ) return d def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=1 , UpperCAmelCase_=False ): if word in self.indices and not overwrite: lowerCAmelCase = self.indices[word] lowerCAmelCase = self.count[idx] + n return idx else: lowerCAmelCase = len(self.symbols ) lowerCAmelCase = idx self.symbols.append(_A ) self.count.append(_A ) return idx def __snake_case ( self , UpperCAmelCase_ ): return 0 def __snake_case ( self , UpperCAmelCase_ ): if isinstance(_A , _A ): try: with open(_A , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(_A ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(_A ) ) return lowerCAmelCase = f.readlines() lowerCAmelCase = self._load_meta(_A ) for line in lines[indices_start_line:]: try: lowerCAmelCase , lowerCAmelCase = line.rstrip().rsplit(''' ''' , 1 ) if field == "#fairseq:overwrite": lowerCAmelCase = True lowerCAmelCase , lowerCAmelCase = line.rsplit(''' ''' , 1 ) else: lowerCAmelCase = False lowerCAmelCase = int(_A ) lowerCAmelCase = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(_A ) ) self.add_symbol(_A , n=_A , overwrite=_A ) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''' ) def UpperCAmelCase ( _snake_case ): # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} lowerCAmelCase = dict((re.sub(R'''@@$''' , '''''' , __snake_case ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , __snake_case ), v) for k, v in d.items() ) lowerCAmelCase = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] lowerCAmelCase = d[k] # restore return da def UpperCAmelCase ( _snake_case , _snake_case ): # prep if not os.path.exists(__snake_case ): raise ValueError(F"""path {biogpt_checkpoint_path} does not exist!""" ) os.makedirs(__snake_case , exist_ok=__snake_case ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models lowerCAmelCase = os.path.join(__snake_case , '''checkpoint.pt''' ) if not os.path.isfile(__snake_case ): raise ValueError(F"""path to the file {checkpoint_file} does not exist!""" ) lowerCAmelCase = torch.load(__snake_case , map_location='''cpu''' ) lowerCAmelCase = chkpt['''cfg''']['''model'''] # dicts lowerCAmelCase = os.path.join(__snake_case , '''dict.txt''' ) if not os.path.isfile(__snake_case ): raise ValueError(F"""path to the file {dict_file} does not exist!""" ) lowerCAmelCase = Dictionary.load(__snake_case ) lowerCAmelCase = rewrite_dict_keys(src_dict.indices ) lowerCAmelCase = len(__snake_case ) lowerCAmelCase = os.path.join(__snake_case , VOCAB_FILES_NAMES['''vocab_file'''] ) print(F"""Generating {src_vocab_file} of {src_vocab_size} records""" ) with open(__snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__snake_case , ensure_ascii=__snake_case , indent=__snake_case ) ) # merges_file (bpecodes) lowerCAmelCase = os.path.join(__snake_case , '''bpecodes''' ) if not os.path.isfile(__snake_case ): raise ValueError(F"""path to the file {bpecodes_file} does not exist!""" ) lowerCAmelCase = os.path.join(__snake_case , VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(__snake_case , __snake_case ) # model config lowerCAmelCase = os.path.join(__snake_case , '''config.json''' ) lowerCAmelCase = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.02, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1E-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(F"""Generating {biogpt_model_config_file}""" ) with open(__snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__snake_case , ensure_ascii=__snake_case , indent=__snake_case ) ) # tokenizer config lowerCAmelCase = os.path.join(__snake_case , __snake_case ) lowerCAmelCase = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 1024, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(F"""Generating {biogpt_tokenizer_config_file}""" ) with open(__snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(__snake_case , ensure_ascii=__snake_case , indent=__snake_case ) ) # model lowerCAmelCase = chkpt['''model'''] # remove unneeded keys lowerCAmelCase = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(__snake_case , __snake_case ) lowerCAmelCase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): lowerCAmelCase = model_state_dict.pop(__snake_case ) else: lowerCAmelCase = model_state_dict.pop(__snake_case ) lowerCAmelCase = BioGptConfig.from_pretrained(__snake_case ) lowerCAmelCase = BioGptForCausalLM(__snake_case ) # check that it loads ok model_new.load_state_dict(__snake_case ) # save lowerCAmelCase = os.path.join(__snake_case , __snake_case ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(__snake_case , __snake_case ) print('''Conversion is done!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() # Required parameters parser.add_argument( """--biogpt_checkpoint_path""", default=None, type=str, required=True, help=( """Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,""" """ bpecodes, etc.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) UpperCAmelCase_ =parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
704
import torch from diffusers import StableDiffusionPipeline UpperCAmelCase_ ="""path-to-your-trained-model""" UpperCAmelCase_ =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") UpperCAmelCase_ ="""A photo of sks dog in a bucket""" UpperCAmelCase_ =pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
33
0
'''simple docstring''' import argparse from collections import defaultdict import yaml UpperCAmelCase_ ='''docs/source/en/_toctree.yml''' def UpperCAmelCase ( _snake_case ): lowerCAmelCase = defaultdict(lowerCAmelCase__ ) lowerCAmelCase = [] lowerCAmelCase = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({'''local''': doc['''local'''], '''title''': doc['''title''']} ) else: new_doc_list.append(lowerCAmelCase__ ) lowerCAmelCase = new_doc_list lowerCAmelCase = [key for key, value in counts.items() if value > 1] lowerCAmelCase = [] for duplicate_key in duplicates: lowerCAmelCase = list({doc['''title'''] for doc in doc_list if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( F"""{duplicate_key} is present several times in the documentation table of content at """ '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if '''local''' not in counts or counts[doc['''local''']] == 1] ) lowerCAmelCase = sorted(lowerCAmelCase__ , key=lambda _snake_case : s["title"].lower() ) # "overview" gets special treatment and is always first if len(lowerCAmelCase__ ) > 1: raise ValueError('''{doc_list} has two \'overview\' docs which is not allowed.''' ) overview_doc.extend(lowerCAmelCase__ ) # Sort return overview_doc def UpperCAmelCase ( _snake_case=False ): with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowerCAmelCase = yaml.safe_load(f.read() ) # Get to the API doc lowerCAmelCase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowerCAmelCase = content[api_idx]['''sections'''] # Then to the model doc lowerCAmelCase = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 lowerCAmelCase = api_doc[scheduler_idx]['''sections'''] lowerCAmelCase = clean_doc_toc(lowerCAmelCase__ ) lowerCAmelCase = False if new_scheduler_doc != scheduler_doc: lowerCAmelCase = True if overwrite: lowerCAmelCase = new_scheduler_doc if diff: if overwrite: lowerCAmelCase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) def UpperCAmelCase ( _snake_case=False ): with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowerCAmelCase = yaml.safe_load(f.read() ) # Get to the API doc lowerCAmelCase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowerCAmelCase = content[api_idx]['''sections'''] # Then to the model doc lowerCAmelCase = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 lowerCAmelCase = False lowerCAmelCase = api_doc[pipeline_idx]['''sections'''] lowerCAmelCase = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: lowerCAmelCase = pipeline_doc['''section'''] lowerCAmelCase = clean_doc_toc(lowerCAmelCase__ ) if overwrite: lowerCAmelCase = new_sub_pipeline_doc new_pipeline_docs.append(lowerCAmelCase__ ) # sort overall pipeline doc lowerCAmelCase = clean_doc_toc(lowerCAmelCase__ ) if new_pipeline_docs != pipeline_docs: lowerCAmelCase = True if overwrite: lowerCAmelCase = new_pipeline_docs if diff: if overwrite: lowerCAmelCase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") UpperCAmelCase_ =parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class __UpperCamelCase ( __lowerCamelCase ): '''simple docstring''' __a : Tuple ='''lilt''' def __init__( self , UpperCAmelCase_=3_05_22 , UpperCAmelCase_=7_68 , UpperCAmelCase_=12 , UpperCAmelCase_=12 , UpperCAmelCase_=30_72 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=5_12 , UpperCAmelCase_=2 , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-1_2 , UpperCAmelCase_=0 , UpperCAmelCase_="absolute" , UpperCAmelCase_=None , UpperCAmelCase_=4 , UpperCAmelCase_=10_24 , **UpperCAmelCase_ , ): super().__init__(pad_token_id=a_ , **a_ ) lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_act lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = initializer_range lowerCAmelCase = layer_norm_eps lowerCAmelCase = position_embedding_type lowerCAmelCase = classifier_dropout lowerCAmelCase = channel_shrink_ratio lowerCAmelCase = max_ad_position_embeddings
706
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = 8 # DPR tok lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __snake_case ( self ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __snake_case ( self ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: lowerCAmelCase = os.path.join(self.tmpdirname , '''dataset''' ) lowerCAmelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCAmelCase_ ) , ) return retriever def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) lowerCAmelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) lowerCAmelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(UpperCAmelCase_ , open(UpperCAmelCase_ , '''wb''' ) ) lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = self.get_dummy_dataset() retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_legacy_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): import torch lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) lowerCAmelCase = retriever( UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ , return_tensors='''pt''' , ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): lowerCAmelCase = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) retriever.set_ctx_encoder_tokenizer(UpperCAmelCase_ ) lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) self.assertEqual( len(UpperCAmelCase_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCAmelCase_ ) # check for doc token related keys in dictionary.
33
0
import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : int ="""naver-clova-ix/donut-base-finetuned-docvqa""" __a : Union[str, Any] =( """This is a tool that answers a question about an document (pdf). It takes an input named `document` which """ """should be the document containing the information, as well as a `question` that is the question about the """ """document. It returns a text that contains the answer to the question.""" ) __a : List[Any] ="""document_qa""" __a : List[Any] =AutoProcessor __a : str =VisionEncoderDecoderModel __a : List[str] =["""image""", """text"""] __a : Union[str, Any] =["""text"""] def __init__( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): if not is_vision_available(): raise ValueError('''Pillow must be installed to use the DocumentQuestionAnsweringTool.''' ) super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' lowerCAmelCase = task_prompt.replace('''{user_input}''' , UpperCAmelCase_ ) lowerCAmelCase = self.pre_processor.tokenizer( UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_tensors='''pt''' ).input_ids lowerCAmelCase = self.pre_processor(UpperCAmelCase_ , return_tensors='''pt''' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def __snake_case ( self , UpperCAmelCase_ ): return self.model.generate( inputs['''pixel_values'''].to(self.device ) , decoder_input_ids=inputs['''decoder_input_ids'''].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=UpperCAmelCase_ , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=UpperCAmelCase_ , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=UpperCAmelCase_ , ).sequences def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.pre_processor.batch_decode(UpperCAmelCase_ )[0] lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '''''' ) lowerCAmelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '''''' ) lowerCAmelCase = re.sub(r'''<.*?>''' , '''''' , UpperCAmelCase_ , count=1 ).strip() # remove first task start token lowerCAmelCase = self.pre_processor.tokenajson(UpperCAmelCase_ ) return sequence["answer"]
707
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """google/switch-base-8""": """https://huggingface.co/google/switch-base-8/blob/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ="""switch_transformers""" __a : Union[str, Any] =["""past_key_values"""] __a : Dict ={"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , UpperCAmelCase_=3_21_28 , UpperCAmelCase_=7_68 , UpperCAmelCase_=64 , UpperCAmelCase_=20_48 , UpperCAmelCase_=64 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=8 , UpperCAmelCase_=False , UpperCAmelCase_=0.01 , UpperCAmelCase_="float32" , UpperCAmelCase_=False , UpperCAmelCase_=32 , UpperCAmelCase_=1_28 , UpperCAmelCase_=0.1 , UpperCAmelCase_=1E-6 , UpperCAmelCase_=0.001 , UpperCAmelCase_=0.001 , UpperCAmelCase_=1.0 , UpperCAmelCase_="relu" , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=0 , UpperCAmelCase_=1 , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = d_kv lowerCAmelCase = d_ff lowerCAmelCase = num_sparse_encoder_layers lowerCAmelCase = num_layers lowerCAmelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: lowerCAmelCase = self.num_layers // self.num_sparse_encoder_layers else: lowerCAmelCase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: lowerCAmelCase = self.num_decoder_layers // self.num_sparse_decoder_layers else: lowerCAmelCase = self.num_decoder_layers # HACK: this will create 0 sparse layers lowerCAmelCase = num_heads lowerCAmelCase = num_experts lowerCAmelCase = expert_capacity lowerCAmelCase = router_bias lowerCAmelCase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) lowerCAmelCase = router_dtype lowerCAmelCase = router_ignore_padding_tokens lowerCAmelCase = relative_attention_num_buckets lowerCAmelCase = relative_attention_max_distance lowerCAmelCase = dropout_rate lowerCAmelCase = layer_norm_epsilon lowerCAmelCase = initializer_factor lowerCAmelCase = feed_forward_proj lowerCAmelCase = use_cache lowerCAmelCase = add_router_probs lowerCAmelCase = router_z_loss_coef lowerCAmelCase = router_aux_loss_coef lowerCAmelCase = self.feed_forward_proj.split('''-''' ) lowerCAmelCase = act_info[-1] lowerCAmelCase = act_info[0] == '''gated''' if len(UpperCAmelCase_ ) > 1 and act_info[0] != "gated" or len(UpperCAmelCase_ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCAmelCase = '''gelu_new''' super().__init__( pad_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ , )
33
0
from typing import Any class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = None class __UpperCamelCase : '''simple docstring''' def __init__( self ): lowerCAmelCase = None def __snake_case ( self ): lowerCAmelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) lowerCAmelCase = temp.next print() def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = Node(_SCREAMING_SNAKE_CASE ) lowerCAmelCase = self.head lowerCAmelCase = new_node def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): if node_data_a == node_data_a: return else: lowerCAmelCase = self.head while node_a is not None and node_a.data != node_data_a: lowerCAmelCase = node_a.next lowerCAmelCase = self.head while node_a is not None and node_a.data != node_data_a: lowerCAmelCase = node_a.next if node_a is None or node_a is None: return lowerCAmelCase , lowerCAmelCase = node_a.data, node_a.data if __name__ == "__main__": UpperCAmelCase_ =LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
708
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def UpperCAmelCase ( _snake_case ): lowerCAmelCase , lowerCAmelCase = analyze_text(_snake_case ) lowerCAmelCase = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase = sum(single_char_strings.values() ) # one length string lowerCAmelCase = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase = single_char_strings[ch] lowerCAmelCase = my_str / all_sum my_fir_sum += prob * math.loga(_snake_case ) # entropy formula. # print entropy print(F"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase = sum(two_char_strings.values() ) lowerCAmelCase = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase = cha + cha if sequence in two_char_strings: lowerCAmelCase = two_char_strings[sequence] lowerCAmelCase = int(_snake_case ) / all_sum my_sec_sum += prob * math.loga(_snake_case ) # print second entropy print(F"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(F"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = Counter() # type: ignore lowerCAmelCase = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(_snake_case ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def UpperCAmelCase ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
33
0
from ...processing_utils import ProcessorMixin class __UpperCamelCase ( _A ): '''simple docstring''' __a : Optional[int] ="""WhisperFeatureExtractor""" __a : Dict ="""WhisperTokenizer""" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ ): super().__init__(UpperCamelCase__ , UpperCamelCase__ ) lowerCAmelCase = self.feature_extractor lowerCAmelCase = False def __snake_case ( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=True ): return self.tokenizer.get_decoder_prompt_ids(task=UpperCamelCase__ , language=UpperCamelCase__ , no_timestamps=UpperCamelCase__ ) def __call__( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): if self._in_target_context_manager: return self.current_processor(*UpperCamelCase__ , **UpperCamelCase__ ) lowerCAmelCase = kwargs.pop('''audio''' , UpperCamelCase__ ) lowerCAmelCase = kwargs.pop('''sampling_rate''' , UpperCamelCase__ ) lowerCAmelCase = kwargs.pop('''text''' , UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: lowerCAmelCase = args[0] lowerCAmelCase = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: lowerCAmelCase = self.feature_extractor(UpperCamelCase__ , *UpperCamelCase__ , sampling_rate=UpperCamelCase__ , **UpperCamelCase__ ) if text is not None: lowerCAmelCase = self.tokenizer(UpperCamelCase__ , **UpperCamelCase__ ) if text is None: return inputs elif audio is None: return encodings else: lowerCAmelCase = encodings['''input_ids'''] return inputs def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.batch_decode(*UpperCamelCase__ , **UpperCamelCase__ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.decode(*UpperCamelCase__ , **UpperCamelCase__ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_="np" ): return self.tokenizer.get_prompt_ids(UpperCamelCase__ , return_tensors=UpperCamelCase__ )
709
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Tuple =IFInpaintingSuperResolutionPipeline __a : Dict =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __a : int =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) __a : Union[str, Any] =PipelineTesterMixin.required_optional_params - {"""latents"""} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
33
0
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed UpperCAmelCase_ ='''true''' def UpperCAmelCase ( _snake_case , _snake_case=82 , _snake_case=16 ): set_seed(42 ) lowerCAmelCase = RegressionModel() lowerCAmelCase = deepcopy(_snake_case ) lowerCAmelCase = RegressionDataset(length=_snake_case ) lowerCAmelCase = DataLoader(_snake_case , batch_size=_snake_case ) model.to(accelerator.device ) lowerCAmelCase = accelerator.prepare(_snake_case , _snake_case ) return model, ddp_model, dataloader def UpperCAmelCase ( _snake_case , _snake_case=False ): lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) lowerCAmelCase = load_dataset('''glue''' , '''mrpc''' , split='''validation''' ) def tokenize_function(_snake_case ): lowerCAmelCase = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=_snake_case , max_length=_snake_case ) return outputs with accelerator.main_process_first(): lowerCAmelCase = dataset.map( _snake_case , batched=_snake_case , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) lowerCAmelCase = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(_snake_case ): if use_longest: return tokenizer.pad(_snake_case , padding='''longest''' , return_tensors='''pt''' ) return tokenizer.pad(_snake_case , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return DataLoader(_snake_case , shuffle=_snake_case , collate_fn=_snake_case , batch_size=16 ) def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = Accelerator(dispatch_batches=_snake_case , split_batches=_snake_case ) lowerCAmelCase = get_dataloader(_snake_case , not dispatch_batches ) lowerCAmelCase = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' , return_dict=_snake_case ) lowerCAmelCase = accelerator.prepare(_snake_case , _snake_case ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [] for batch in dataloader: lowerCAmelCase = batch.values() with torch.no_grad(): lowerCAmelCase = model(_snake_case ) lowerCAmelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) lowerCAmelCase = [], [] for logit, targ in logits_and_targets: logits.append(_snake_case ) targs.append(_snake_case ) lowerCAmelCase = torch.cat(_snake_case ), torch.cat(_snake_case ) return logits, targs def UpperCAmelCase ( _snake_case , _snake_case=82 , _snake_case=False , _snake_case=False , _snake_case=16 ): lowerCAmelCase = get_basic_setup(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = generate_predictions(_snake_case , _snake_case , _snake_case ) assert ( len(_snake_case ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_snake_case )}""" def UpperCAmelCase ( _snake_case = False , _snake_case = False ): lowerCAmelCase = evaluate.load('''glue''' , '''mrpc''' ) lowerCAmelCase = get_mrpc_setup(_snake_case , _snake_case ) # First do baseline lowerCAmelCase = setup['no'] model.to(_snake_case ) model.eval() for batch in dataloader: batch.to(_snake_case ) with torch.inference_mode(): lowerCAmelCase = model(**_snake_case ) lowerCAmelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_snake_case , references=batch['''labels'''] ) lowerCAmelCase = metric.compute() # Then do distributed lowerCAmelCase = setup['ddp'] model.eval() for batch in dataloader: with torch.inference_mode(): lowerCAmelCase = model(**_snake_case ) lowerCAmelCase = outputs.logits.argmax(dim=-1 ) lowerCAmelCase = batch['labels'] lowerCAmelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_snake_case , references=_snake_case ) lowerCAmelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def UpperCAmelCase ( ): lowerCAmelCase = Accelerator(split_batches=_snake_case , dispatch_batches=_snake_case ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(_snake_case , _snake_case ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: lowerCAmelCase = Accelerator(split_batches=_snake_case , dispatch_batches=_snake_case ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(_snake_case , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) lowerCAmelCase = Accelerator() test_torch_metrics(_snake_case , 512 ) accelerator.state._reset_state() def UpperCAmelCase ( _snake_case ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
710
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ ={ """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=7 , UpperCAmelCase_=6 , UpperCAmelCase_=17 , UpperCAmelCase_=23 , UpperCAmelCase_=11 , UpperCAmelCase_=True , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = act_dim lowerCAmelCase = state_dim lowerCAmelCase = hidden_size lowerCAmelCase = max_length lowerCAmelCase = is_training def __snake_case ( self ): lowerCAmelCase = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowerCAmelCase = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowerCAmelCase = floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCAmelCase = floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCAmelCase = ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowerCAmelCase = random_attention_mask((self.batch_size, self.seq_length) ) lowerCAmelCase = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def __snake_case ( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ): lowerCAmelCase = DecisionTransformerModel(config=__a ) model.to(__a ) model.eval() lowerCAmelCase = model(__a , __a , __a , __a , __a , __a ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = { '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : List[Any] =(DecisionTransformerModel,) if is_torch_available() else () __a : Any =() __a : List[Any] ={"""feature-extraction""": DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids __a : Dict =False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features __a : int =False __a : List[Any] =False __a : List[Any] =False __a : int =False __a : Dict =False __a : Any =False __a : Dict =False __a : List[str] =False __a : str =False def __snake_case ( self ): lowerCAmelCase = DecisionTransformerModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=__a , hidden_size=37 ) def __snake_case ( self ): self.config_tester.run_common_tests() def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) @slow def __snake_case ( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = DecisionTransformerModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(__a ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = [ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(__a )] , __a ) @require_torch class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __snake_case ( self ): lowerCAmelCase = 2 # number of steps of autoregressive prediction we will perform lowerCAmelCase = 10 # defined by the RL environment, may be normalized lowerCAmelCase = DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowerCAmelCase = model.to(__a ) lowerCAmelCase = model.config torch.manual_seed(0 ) lowerCAmelCase = torch.randn(1 , 1 , config.state_dim ).to(device=__a , dtype=torch.floataa ) # env.reset() lowerCAmelCase = torch.tensor( [[0.242793, -0.28693074, 0.8742613], [0.67815274, -0.08101085, -0.12952147]] , device=__a ) lowerCAmelCase = torch.tensor(__a , device=__a , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowerCAmelCase = state lowerCAmelCase = torch.zeros(1 , 0 , config.act_dim , device=__a , dtype=torch.floataa ) lowerCAmelCase = torch.zeros(1 , 0 , device=__a , dtype=torch.floataa ) lowerCAmelCase = torch.tensor(0 , device=__a , dtype=torch.long ).reshape(1 , 1 ) for step in range(__a ): lowerCAmelCase = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=__a )] , dim=1 ) lowerCAmelCase = torch.cat([rewards, torch.zeros(1 , 1 , device=__a )] , dim=1 ) lowerCAmelCase = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = model( states=__a , actions=__a , rewards=__a , returns_to_go=__a , timesteps=__a , attention_mask=__a , return_dict=__a , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=__a , dtype=torch.floataa ), 1.0, False, {}, ) lowerCAmelCase = action_pred[0, -1] lowerCAmelCase = torch.cat([states, state] , dim=1 ) lowerCAmelCase = returns_to_go[0, -1] - reward lowerCAmelCase = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowerCAmelCase = torch.cat( [timesteps, torch.ones((1, 1) , device=__a , dtype=torch.long ) * (step + 1)] , dim=1 )
711
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline UpperCAmelCase_ =datasets.utils.logging.get_logger(__name__) @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __a : Optional[datasets.Features] =None __a : str ="utf-8" __a : Optional[str] =None __a : Optional[str] =None __a : bool =True # deprecated __a : Optional[int] =None # deprecated __a : int =1_0 << 2_0 # 10MB __a : Optional[bool] =None class __UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __a : str =JsonConfig def __snake_case ( self ): if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowerCAmelCase = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def __snake_case ( self , UpperCAmelCase_ ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase_ , (str, list, tuple) ): lowerCAmelCase = data_files if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase_ , gen_kwargs={'''files''': files} ) ) return splits def __snake_case ( self , UpperCAmelCase_ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCAmelCase = self.config.features.arrow_schema.field(UpperCAmelCase_ ).type lowerCAmelCase = pa_table.append_column(UpperCAmelCase_ , pa.array([None] * len(UpperCAmelCase_ ) , type=UpperCAmelCase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase = table_cast(UpperCAmelCase_ , self.config.features.arrow_schema ) return pa_table def __snake_case ( self , UpperCAmelCase_ ): for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) # We keep only the field we are interested in lowerCAmelCase = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(UpperCAmelCase_ , (list, tuple) ): lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} else: lowerCAmelCase = dataset lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) yield file_idx, self._cast_table(UpperCAmelCase_ ) # If the file has one json object per line else: with open(UpperCAmelCase_ , '''rb''' ) as f: lowerCAmelCase = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCAmelCase = max(self.config.chunksize // 32 , 16 << 10 ) lowerCAmelCase = ( self.config.encoding_errors if self.config.encoding_errors is not None else '''strict''' ) while True: lowerCAmelCase = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(UpperCAmelCase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCAmelCase = batch.decode(self.config.encoding , errors=UpperCAmelCase_ ).encode('''utf-8''' ) try: while True: try: lowerCAmelCase = paj.read_json( io.BytesIO(UpperCAmelCase_ ) , read_options=paj.ReadOptions(block_size=UpperCAmelCase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(UpperCAmelCase_ , pa.ArrowInvalid ) and "straddling" not in str(UpperCAmelCase_ ) or block_size > len(UpperCAmelCase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(UpperCAmelCase_ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # list is the only sequence type supported in JSON try: lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(UpperCAmelCase_ ) break else: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase_ ) batch_idx += 1
33
0
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( lowercase_ , unittest.TestCase ): '''simple docstring''' __a : Any =CLIPTokenizer __a : str =CLIPTokenizerFast __a : int =True __a : Tuple ={} __a : Any =False def __snake_case ( self ): super().setUp() # fmt: off lowerCAmelCase = ["""l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on lowerCAmelCase = dict(zip(lowerCamelCase_ , range(len(lowerCamelCase_ ) ) ) ) lowerCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>"""] lowerCAmelCase = {"""unk_token""": """<unk>"""} lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(lowerCamelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(lowerCamelCase_ ) ) def __snake_case ( self , **UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase_ ) def __snake_case ( self , **UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = """lower newer""" lowerCAmelCase = """lower newer""" return input_text, output_text def __snake_case ( self ): lowerCAmelCase = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCAmelCase = """lower newer""" lowerCAmelCase = ["""lo""", """w""", """er</w>""", """n""", """e""", """w""", """er</w>"""] lowerCAmelCase = tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) lowerCAmelCase = tokens + [tokenizer.unk_token] lowerCAmelCase = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , lowerCamelCase_ ) @require_ftfy def __snake_case ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = self.tokenizer_class.from_pretrained(lowerCamelCase_ , **lowerCamelCase_ ) lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase_ , **lowerCamelCase_ ) lowerCAmelCase = """A\n'll 11p223RF☆ho!!to?'d'd''d of a cat to-$''d.""" lowerCAmelCase = tokenizer_s.tokenize(lowerCamelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways lowerCAmelCase = """xa\u0303y""" + """ """ + """x\xe3y""" lowerCAmelCase = tokenizer_s.tokenize(lowerCamelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) # Test that the tokenization is identical on unicode of space type lowerCAmelCase = [ """\u0009""", # (horizontal tab, '\t') """\u000B""", # (vertical tab) """\u000C""", # (form feed) """\u0020""", # (space, ' ') """\u200E""", # (left-to-right mark):w """\u200F""", # (right-to-left mark) ] for unicode_seq in spaces_unicodes: lowerCAmelCase = tokenizer_s.tokenize(lowerCamelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) # Test that the tokenization is identical on unicode of line break type lowerCAmelCase = [ """\u000A""", # (line feed, '\n') """\r\n""", # (carriage return and line feed, '\r\n') """\u000D""", # (carriage return, '\r') """\r""", # (carriage return, '\r') """\u000D""", # (carriage return, '\r') """\u2028""", # (line separator) """\u2029""", # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: lowerCAmelCase = tokenizer_s.tokenize(lowerCamelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) def __snake_case ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` lowerCAmelCase = F"""{text_of_1_token} {text_of_1_token}""" lowerCAmelCase = self.rust_tokenizer_class.from_pretrained( lowerCamelCase_ , use_fast=lowerCamelCase_ , ) lowerCAmelCase = tokenizer_r(lowerCamelCase_ , return_offsets_mapping=lowerCamelCase_ , add_special_tokens=lowerCamelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCamelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCamelCase_ ) + 1, len(lowerCamelCase_ ) + 1 + len(lowerCamelCase_ )) , ) lowerCAmelCase = F""" {text}""" lowerCAmelCase = self.rust_tokenizer_class.from_pretrained( lowerCamelCase_ , use_fast=lowerCamelCase_ , ) lowerCAmelCase = tokenizer_r(lowerCamelCase_ , return_offsets_mapping=lowerCamelCase_ , add_special_tokens=lowerCamelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(lowerCamelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCamelCase_ ) + 1, 1 + len(lowerCamelCase_ ) + 1 + len(lowerCamelCase_ )) , ) def __snake_case ( self ): with self.assertRaises(lowerCamelCase_ ) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''' ) self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''' ) ) @require_ftfy def __snake_case ( self ): super().test_tokenization_python_rust_equals() def __snake_case ( self ): pass
712
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] ="""maskformer-swin""" __a : Optional[int] ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , UpperCAmelCase_=2_24 , UpperCAmelCase_=4 , UpperCAmelCase_=3 , UpperCAmelCase_=96 , UpperCAmelCase_=[2, 2, 6, 2] , UpperCAmelCase_=[3, 6, 12, 24] , UpperCAmelCase_=7 , UpperCAmelCase_=4.0 , UpperCAmelCase_=True , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.1 , UpperCAmelCase_="gelu" , UpperCAmelCase_=False , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-5 , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = embed_dim lowerCAmelCase = depths lowerCAmelCase = len(UpperCAmelCase_ ) lowerCAmelCase = num_heads lowerCAmelCase = window_size lowerCAmelCase = mlp_ratio lowerCAmelCase = qkv_bias lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = drop_path_rate lowerCAmelCase = hidden_act lowerCAmelCase = use_absolute_embeddings lowerCAmelCase = layer_norm_eps lowerCAmelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase = int(embed_dim * 2 ** (len(UpperCAmelCase_ ) - 1) ) lowerCAmelCase = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase_ ) + 1 )] lowerCAmelCase , lowerCAmelCase = get_aligned_output_features_output_indices( out_features=UpperCAmelCase_ , out_indices=UpperCAmelCase_ , stage_names=self.stage_names )
33
0
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCamelCase ( __lowerCamelCase ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): return 0.0 def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) lowerCAmelCase = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = 512 lowerCAmelCase = [1] + [0] * (size - 1) lowerCAmelCase = [filter_type.process(A__ ) for item in inputs] lowerCAmelCase = [0] * (samplerate - size) # zero-padding outputs += filler lowerCAmelCase = np.abs(np.fft.fft(A__ ) ) lowerCAmelCase = 20 * np.logaa(A__ ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) # Display within reasonable bounds lowerCAmelCase = get_bounds(A__ , A__ ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('''Gain (dB)''' ) plt.plot(A__ ) plt.show() def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = 512 lowerCAmelCase = [1] + [0] * (size - 1) lowerCAmelCase = [filter_type.process(A__ ) for item in inputs] lowerCAmelCase = [0] * (samplerate - size) # zero-padding outputs += filler lowerCAmelCase = np.angle(np.fft.fft(A__ ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('''Phase shift (Radians)''' ) plt.plot(np.unwrap(A__ , -2 * pi ) ) plt.show()
713
from collections.abc import Sequence def UpperCAmelCase ( _snake_case , _snake_case = False ): if not arr: return 0 lowerCAmelCase = 0 if allow_empty_subarrays else float('''-inf''' ) lowerCAmelCase = 0.0 for num in arr: lowerCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) lowerCAmelCase = max(_snake_case , _snake_case ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase_ =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'''{max_subarray_sum(nums) = }''')
33
0
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class __UpperCamelCase ( _UpperCamelCase ): '''simple docstring''' __a : Any ="""informer""" __a : Optional[Any] ={ """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = "student_t" , UpperCAmelCase_ = "nll" , UpperCAmelCase_ = 1 , UpperCAmelCase_ = None , UpperCAmelCase_ = "mean" , UpperCAmelCase_ = 0 , UpperCAmelCase_ = 0 , UpperCAmelCase_ = 0 , UpperCAmelCase_ = 0 , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = 64 , UpperCAmelCase_ = 32 , UpperCAmelCase_ = 32 , UpperCAmelCase_ = 2 , UpperCAmelCase_ = 2 , UpperCAmelCase_ = 2 , UpperCAmelCase_ = 2 , UpperCAmelCase_ = True , UpperCAmelCase_ = "gelu" , UpperCAmelCase_ = 0.05 , UpperCAmelCase_ = 0.1 , UpperCAmelCase_ = 0.1 , UpperCAmelCase_ = 0.1 , UpperCAmelCase_ = 0.1 , UpperCAmelCase_ = 1_00 , UpperCAmelCase_ = 0.02 , UpperCAmelCase_=True , UpperCAmelCase_ = "prob" , UpperCAmelCase_ = 5 , UpperCAmelCase_ = True , **UpperCAmelCase_ , ): # time series specific configuration lowerCAmelCase = prediction_length lowerCAmelCase = context_length or prediction_length lowerCAmelCase = distribution_output lowerCAmelCase = loss lowerCAmelCase = input_size lowerCAmelCase = num_time_features lowerCAmelCase = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] lowerCAmelCase = scaling lowerCAmelCase = num_dynamic_real_features lowerCAmelCase = num_static_real_features lowerCAmelCase = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(UpperCAmelCase_ ) != num_static_categorical_features: raise ValueError( '''The cardinality should be a list of the same length as `num_static_categorical_features`''' ) lowerCAmelCase = cardinality else: lowerCAmelCase = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(UpperCAmelCase_ ) != num_static_categorical_features: raise ValueError( '''The embedding dimension should be a list of the same length as `num_static_categorical_features`''' ) lowerCAmelCase = embedding_dimension else: lowerCAmelCase = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCAmelCase = num_parallel_samples # Transformer architecture configuration lowerCAmelCase = input_size * len(self.lags_sequence ) + self._number_of_features lowerCAmelCase = d_model lowerCAmelCase = encoder_attention_heads lowerCAmelCase = decoder_attention_heads lowerCAmelCase = encoder_ffn_dim lowerCAmelCase = decoder_ffn_dim lowerCAmelCase = encoder_layers lowerCAmelCase = decoder_layers lowerCAmelCase = dropout lowerCAmelCase = attention_dropout lowerCAmelCase = activation_dropout lowerCAmelCase = encoder_layerdrop lowerCAmelCase = decoder_layerdrop lowerCAmelCase = activation_function lowerCAmelCase = init_std lowerCAmelCase = use_cache # Informer lowerCAmelCase = attention_type lowerCAmelCase = sampling_factor lowerCAmelCase = distil super().__init__(is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ ) @property def __snake_case ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
714
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =BertJapaneseTokenizer __a : Optional[int] =False __a : int =True def __snake_case ( self ): super().setUp() lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(do_lower_case=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer( do_lower_case=UpperCAmelCase_ , normalize_text=UpperCAmelCase_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(normalize_text=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(do_lower_case=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(normalize_text=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(trim_whitespace=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(do_lower_case=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(normalize_text=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(trim_whitespace=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = WordpieceTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) lowerCAmelCase = tokenizer.subword_tokenizer lowerCAmelCase = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) lowerCAmelCase = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =BertJapaneseTokenizer __a : Optional[int] =False def __snake_case ( self ): super().setUp() lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , **UpperCAmelCase_ ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( UpperCAmelCase_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = CharacterTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) lowerCAmelCase = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
33
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = 3 lowerCAmelCase = (32, 32) lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowercase_ ) return image @property def __snake_case ( self ): torch.manual_seed(0 ) lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=lowercase_ , only_cross_attention=(True, True, False) , num_class_embeds=1_00 , ) return model @property def __snake_case ( self ): torch.manual_seed(0 ) lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def __snake_case ( self ): torch.manual_seed(0 ) lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=5_12 , ) return CLIPTextModel(lowercase_ ) def __snake_case ( self ): lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase = self.dummy_cond_unet_upscale lowerCAmelCase = DDPMScheduler() lowerCAmelCase = DDIMScheduler(prediction_type='''v_prediction''' ) lowerCAmelCase = self.dummy_vae lowerCAmelCase = self.dummy_text_encoder lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(lowercase_ ) ).convert('''RGB''' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk lowerCAmelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=3_50 , ) lowerCAmelCase = sd_pipe.to(lowercase_ ) sd_pipe.set_progress_bar_config(disable=lowercase_ ) lowerCAmelCase = '''A painting of a squirrel eating a burger''' lowerCAmelCase = torch.Generator(device=lowercase_ ).manual_seed(0 ) lowerCAmelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = torch.Generator(device=lowercase_ ).manual_seed(0 ) lowerCAmelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , return_dict=lowercase_ , )[0] lowerCAmelCase = image[0, -3:, -3:, -1] lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] lowerCAmelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) lowerCAmelCase = np.array([0.3113, 0.3910, 0.4272, 0.4859, 0.5061, 0.4652, 0.5362, 0.5715, 0.5661] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __snake_case ( self ): lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase = self.dummy_cond_unet_upscale lowerCAmelCase = DDPMScheduler() lowerCAmelCase = DDIMScheduler(prediction_type='''v_prediction''' ) lowerCAmelCase = self.dummy_vae lowerCAmelCase = self.dummy_text_encoder lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(lowercase_ ) ).convert('''RGB''' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk lowerCAmelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=3_50 , ) lowerCAmelCase = sd_pipe.to(lowercase_ ) sd_pipe.set_progress_bar_config(disable=lowercase_ ) lowerCAmelCase = '''A painting of a squirrel eating a burger''' lowerCAmelCase = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) lowerCAmelCase = output.images assert image.shape[0] == 2 lowerCAmelCase = torch.Generator(device=lowercase_ ).manual_seed(0 ) lowerCAmelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) lowerCAmelCase = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def __snake_case ( self ): lowerCAmelCase = self.dummy_cond_unet_upscale lowerCAmelCase = DDPMScheduler() lowerCAmelCase = DDIMScheduler(prediction_type='''v_prediction''' ) lowerCAmelCase = self.dummy_vae lowerCAmelCase = self.dummy_text_encoder lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(lowercase_ ) ).convert('''RGB''' ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 lowerCAmelCase = unet.half() lowerCAmelCase = text_encoder.half() # make sure here that pndm scheduler skips prk lowerCAmelCase = StableDiffusionUpscalePipeline( unet=lowercase_ , low_res_scheduler=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , max_noise_level=3_50 , ) lowerCAmelCase = sd_pipe.to(lowercase_ ) sd_pipe.set_progress_bar_config(disable=lowercase_ ) lowerCAmelCase = '''A painting of a squirrel eating a burger''' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = sd_pipe( [prompt] , image=lowercase_ , generator=lowercase_ , num_inference_steps=2 , output_type='''np''' , ).images lowerCAmelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) lowerCAmelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat.npy''' ) lowerCAmelCase = '''stabilityai/stable-diffusion-x4-upscaler''' lowerCAmelCase = StableDiffusionUpscalePipeline.from_pretrained(lowercase_ ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) pipe.enable_attention_slicing() lowerCAmelCase = '''a cat sitting on a park bench''' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , output_type='''np''' , ) lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 1E-3 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) lowerCAmelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat_fp16.npy''' ) lowerCAmelCase = '''stabilityai/stable-diffusion-x4-upscaler''' lowerCAmelCase = StableDiffusionUpscalePipeline.from_pretrained( lowercase_ , torch_dtype=torch.floataa , ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) pipe.enable_attention_slicing() lowerCAmelCase = '''a cat sitting on a park bench''' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , output_type='''np''' , ) lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 5E-1 def __snake_case ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) lowerCAmelCase = '''stabilityai/stable-diffusion-x4-upscaler''' lowerCAmelCase = StableDiffusionUpscalePipeline.from_pretrained( lowercase_ , torch_dtype=torch.floataa , ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCAmelCase = '''a cat sitting on a park bench''' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe( prompt=lowercase_ , image=lowercase_ , generator=lowercase_ , num_inference_steps=5 , output_type='''np''' , ) lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
715
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) UpperCAmelCase_ ="""hf-internal-testing/tiny-random-bert""" UpperCAmelCase_ =os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") UpperCAmelCase_ ="""9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCAmelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertTrue(os.path.isfile(UpperCAmelCase_ ) ) # File is cached at the same place the second time. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Using a specific revision to test the full commit hash. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''9b8c223''' ) self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): lowerCAmelCase = cached_file('''tiny-random-bert''' , UpperCAmelCase_ ) with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCAmelCase_ , '''.no_exist''' , UpperCAmelCase_ , '''conf''' ) ) ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , local_files_only=UpperCAmelCase_ , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = mock.Mock() lowerCAmelCase = 5_00 lowerCAmelCase = {} lowerCAmelCase = HTTPError lowerCAmelCase = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCAmelCase_ ) as mock_head: lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) # This check we did call the fake head request mock_head.assert_called() def __snake_case ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) def __snake_case ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCAmelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ , revision='''ahaha''' ) lowerCAmelCase = get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCAmelCase = json.loads(open(UpperCAmelCase_ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_68 ) def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase = Path(UpperCAmelCase_ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCAmelCase_ , '''a.txt''' ) , str(UpperCAmelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCAmelCase_ , '''b.txt''' ) )
33
0
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) UpperCAmelCase_ =logging.getLogger(__name__) @dataclass(frozen=__lowerCAmelCase ) class __UpperCamelCase : '''simple docstring''' __a : str __a : str __a : Optional[str] =None __a : Optional[str] =None __a : Optional[str] =None @dataclass(frozen=__lowerCAmelCase ) class __UpperCamelCase : '''simple docstring''' __a : List[int] __a : Optional[List[int]] =None __a : Optional[List[int]] =None __a : Optional[Union[int, float]] =None __a : Optional[int] =None if is_torch_available(): import torch from torch.utils.data import Dataset class __UpperCamelCase ( __lowerCAmelCase ): '''simple docstring''' __a : List[InputFeatures] def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_=False , UpperCAmelCase_ = False , ): lowerCAmelCase = hans_processors[task]() lowerCAmelCase = os.path.join( lowerCAmelCase_ , '''cached_{}_{}_{}_{}'''.format( '''dev''' if evaluate else '''train''' , tokenizer.__class__.__name__ , str(lowerCAmelCase_ ) , lowerCAmelCase_ , ) , ) lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase = cached_features_file + '''.lock''' with FileLock(lowerCAmelCase_ ): if os.path.exists(lowerCAmelCase_ ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) lowerCAmelCase = torch.load(lowerCAmelCase_ ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) lowerCAmelCase = ( processor.get_dev_examples(lowerCAmelCase_ ) if evaluate else processor.get_train_examples(lowerCAmelCase_ ) ) logger.info('''Training examples: %s''' , len(lowerCAmelCase_ ) ) lowerCAmelCase = hans_convert_examples_to_features(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) logger.info('''Saving features into cached file %s''' , lowerCAmelCase_ ) torch.save(self.features , lowerCAmelCase_ ) def __len__( self ): return len(self.features ) def __getitem__( self , UpperCAmelCase_ ): return self.features[i] def __snake_case ( self ): return self.label_list if is_tf_available(): import tensorflow as tf class __UpperCamelCase : '''simple docstring''' __a : List[InputFeatures] def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 1_28 , UpperCAmelCase_=False , UpperCAmelCase_ = False , ): lowerCAmelCase = hans_processors[task]() lowerCAmelCase = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase , lowerCAmelCase = label_list[2], label_list[1] lowerCAmelCase = label_list lowerCAmelCase = processor.get_dev_examples(lowerCAmelCase_ ) if evaluate else processor.get_train_examples(lowerCAmelCase_ ) lowerCAmelCase = hans_convert_examples_to_features(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='''convert examples to features''' ): if ex_index % 1_00_00 == 0: logger.info('''Writing example %d of %d''' % (ex_index, len(lowerCAmelCase_ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCAmelCase = tf.data.Dataset.from_generator( lowerCAmelCase_ , ( { '''example_id''': tf.intaa, '''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa, }, tf.intaa, ) , ( { '''example_id''': tf.TensorShape([] ), '''input_ids''': tf.TensorShape([None, None] ), '''attention_mask''': tf.TensorShape([None, None] ), '''token_type_ids''': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def __snake_case ( self ): return self.dataset def __len__( self ): return len(self.features ) def __getitem__( self , UpperCAmelCase_ ): return self.features[i] def __snake_case ( self ): return self.label_list class __UpperCamelCase ( __lowerCAmelCase ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): return self._create_examples(self._read_tsv(os.path.join(lowerCAmelCase_ , '''heuristics_train_set.txt''' ) ) , '''train''' ) def __snake_case ( self , UpperCAmelCase_ ): return self._create_examples(self._read_tsv(os.path.join(lowerCAmelCase_ , '''heuristics_evaluation_set.txt''' ) ) , '''dev''' ) def __snake_case ( self ): return ["contradiction", "entailment", "neutral"] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [] for i, line in enumerate(lowerCAmelCase_ ): if i == 0: continue lowerCAmelCase = '''%s-%s''' % (set_type, line[0]) lowerCAmelCase = line[5] lowerCAmelCase = line[6] lowerCAmelCase = line[7][2:] if line[7].startswith('''ex''' ) else line[7] lowerCAmelCase = line[0] examples.append(InputExample(guid=lowerCAmelCase_ , text_a=lowerCAmelCase_ , text_b=lowerCAmelCase_ , label=lowerCAmelCase_ , pairID=lowerCAmelCase_ ) ) return examples def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case , ): lowerCAmelCase = {label: i for i, label in enumerate(snake_case__ )} lowerCAmelCase = [] for ex_index, example in tqdm.tqdm(enumerate(snake_case__ ) , desc='''convert examples to features''' ): if ex_index % 10000 == 0: logger.info('''Writing example %d''' % (ex_index) ) lowerCAmelCase = tokenizer( example.text_a , example.text_b , add_special_tokens=snake_case__ , max_length=snake_case__ , padding='''max_length''' , truncation=snake_case__ , return_overflowing_tokens=snake_case__ , ) lowerCAmelCase = label_map[example.label] if example.label in label_map else 0 lowerCAmelCase = int(example.pairID ) features.append(InputFeatures(**snake_case__ , label=snake_case__ , pairID=snake_case__ ) ) for i, example in enumerate(examples[:5] ): logger.info('''*** Example ***''' ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features UpperCAmelCase_ ={ """hans""": 3, } UpperCAmelCase_ ={ """hans""": HansProcessor, }
716
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = "arrow" , **UpperCAmelCase_ , ): super().__init__( split=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , keep_in_memory=UpperCAmelCase_ , streaming=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = load_from_cache_file lowerCAmelCase = file_format lowerCAmelCase = Spark( df=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , working_dir=UpperCAmelCase_ , **UpperCAmelCase_ , ) def __snake_case ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
33
0
import re from filelock import FileLock try: import nltk UpperCAmelCase_ =True except (ImportError, ModuleNotFoundError): UpperCAmelCase_ =False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def UpperCAmelCase ( _snake_case ): re.sub('''<n>''' , '''''' , _snake_case ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(_snake_case ) )
717
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase ( _snake_case = 3 ): if isinstance(_snake_case , _snake_case ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(_snake_case ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) lowerCAmelCase = QuantumRegister(_snake_case , '''qr''' ) lowerCAmelCase = ClassicalRegister(_snake_case , '''cr''' ) lowerCAmelCase = QuantumCircuit(_snake_case , _snake_case ) lowerCAmelCase = number_of_qubits for i in range(_snake_case ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(_snake_case ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , _snake_case , _snake_case ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(_snake_case , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(_snake_case , _snake_case ) # simulate with 10000 shots lowerCAmelCase = Aer.get_backend('''qasm_simulator''' ) lowerCAmelCase = execute(_snake_case , _snake_case , shots=10000 ) return job.result().get_counts(_snake_case ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
33
0
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration UpperCAmelCase_ =HfArgumentParser(InitializationArguments) UpperCAmelCase_ =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization UpperCAmelCase_ =AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks UpperCAmelCase_ ={ """vocab_size""": len(tokenizer), """scale_attn_by_inverse_layer_idx""": True, """reorder_and_upcast_attn""": True, } # Load model config (GPT-2 large in this case) UpperCAmelCase_ =AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config UpperCAmelCase_ =AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
718
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Any =1 @register_to_config def __init__( self , UpperCAmelCase_=20_00 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=1E-3 ): lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = torch.linspace(1 , self.config.sampling_eps , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCAmelCase = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCAmelCase = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCAmelCase = std.flatten() while len(std.shape ) < len(score.shape ): lowerCAmelCase = std.unsqueeze(-1 ) lowerCAmelCase = -score / std # compute lowerCAmelCase = -1.0 / len(self.timesteps ) lowerCAmelCase = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCAmelCase = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCAmelCase = beta_t.unsqueeze(-1 ) lowerCAmelCase = -0.5 * beta_t * x lowerCAmelCase = torch.sqrt(UpperCAmelCase_ ) lowerCAmelCase = drift - diffusion**2 * score lowerCAmelCase = x + drift * dt # add noise lowerCAmelCase = randn_tensor(x.shape , layout=x.layout , generator=UpperCAmelCase_ , device=x.device , dtype=x.dtype ) lowerCAmelCase = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ): return self.config.num_train_timesteps
33
0
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path UpperCAmelCase_ =Path(__file__).resolve().parents[3] / """src""" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) UpperCAmelCase_ ={"""base""": """patrickvonplaten/wav2vec2_tiny_random""", """robust""": """patrickvonplaten/wav2vec2_tiny_random_robust"""} UpperCAmelCase_ ="""zero2""" UpperCAmelCase_ ="""zero3""" UpperCAmelCase_ =[ZEROa, ZEROa] def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = parameterized.to_safe_name('''_'''.join(str(__a ) for x in param.args ) ) return F"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test UpperCAmelCase_ =list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class __UpperCamelCase ( __lowerCamelCase ): '''simple docstring''' @parameterized.expand(SCREAMING_SNAKE_CASE_ , name_func=SCREAMING_SNAKE_CASE_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ , model=SCREAMING_SNAKE_CASE_ , distributed=SCREAMING_SNAKE_CASE_ , fpaa=SCREAMING_SNAKE_CASE_ , ) @require_torch_multi_gpu @parameterized.expand(SCREAMING_SNAKE_CASE_ , name_func=SCREAMING_SNAKE_CASE_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ , model=SCREAMING_SNAKE_CASE_ , distributed=SCREAMING_SNAKE_CASE_ , fpaa=SCREAMING_SNAKE_CASE_ , ) @parameterized.expand(SCREAMING_SNAKE_CASE_ , name_func=SCREAMING_SNAKE_CASE_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ , model=SCREAMING_SNAKE_CASE_ , distributed=SCREAMING_SNAKE_CASE_ , fpaa=SCREAMING_SNAKE_CASE_ , ) @require_torch_multi_gpu @parameterized.expand(SCREAMING_SNAKE_CASE_ , name_func=SCREAMING_SNAKE_CASE_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ , model=SCREAMING_SNAKE_CASE_ , distributed=SCREAMING_SNAKE_CASE_ , fpaa=SCREAMING_SNAKE_CASE_ , ) def __snake_case ( self , UpperCAmelCase_ ): # XXX: run_asr is premature and doesn't save any results # so all we check for now is that the process didn't fail pass def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 10 , UpperCAmelCase_ = True , UpperCAmelCase_ = True , UpperCAmelCase_ = True , ): lowerCAmelCase = models[model] lowerCAmelCase = self.run_trainer( stage=SCREAMING_SNAKE_CASE_ , model_name=SCREAMING_SNAKE_CASE_ , eval_steps=SCREAMING_SNAKE_CASE_ , num_train_epochs=1 , distributed=SCREAMING_SNAKE_CASE_ , fpaa=SCREAMING_SNAKE_CASE_ , ) self.do_checks(SCREAMING_SNAKE_CASE_ ) return output_dir def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 10 , UpperCAmelCase_ = 1 , UpperCAmelCase_ = True , UpperCAmelCase_ = True , ): lowerCAmelCase = self.get_auto_remove_tmp_dir('''./xxx''' , after=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase = F"""\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(SCREAMING_SNAKE_CASE_ )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n """.split() if fpaa: args.extend(['''--fp16'''] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files lowerCAmelCase = F"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() lowerCAmelCase = [F"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] lowerCAmelCase = self.get_launcher(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(SCREAMING_SNAKE_CASE_ , env=self.get_env() ) return output_dir def __snake_case ( self , UpperCAmelCase_=False ): # 1. explicitly set --num_nodes=1 just in case these tests end up run on a multi-node setup # - it won't be able to handle that # 2. for now testing with just 2 gpus max (since some quality tests may give different # results with mode gpus because we use very little data) lowerCAmelCase = min(2 , get_gpu_count() ) if distributed else 1 return F"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
719
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
33
0
def UpperCAmelCase ( _snake_case ): if not isinstance(_snake_case , _snake_case ): raise TypeError('''Input value must be an \'int\' type''' ) lowerCAmelCase = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
720
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : str =MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = VideoClassificationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ , top_k=2 ) lowerCAmelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): for example in examples: lowerCAmelCase = video_classifier(UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, ] , ) @require_torch def __snake_case ( self ): lowerCAmelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' lowerCAmelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) lowerCAmelCase = pipeline( '''video-classification''' , model=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = video_classifier(UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __snake_case ( self ): pass
33
0
import math def UpperCAmelCase ( _snake_case , _snake_case ): if ( not isinstance(_snake_case , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * power_factor def UpperCAmelCase ( _snake_case , _snake_case ): if ( not isinstance(_snake_case , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
721
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any ="""hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __snake_case ( self , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCAmelCase_ ) ) lowerCAmelCase = np.random.RandomState(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # warmup pass to apply optimizations lowerCAmelCase = pipe(**self.get_dummy_inputs() ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @property def __snake_case ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __snake_case ( self ): lowerCAmelCase = ort.SessionOptions() lowerCAmelCase = False return options def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCAmelCase_ , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
33
0
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): # Initialise PyTorch model lowerCAmelCase = TaConfig.from_json_file(_lowerCAmelCase ) print(F"""Building PyTorch model from configuration: {config}""" ) lowerCAmelCase = TaForConditionalGeneration(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_ta(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) UpperCAmelCase_ =parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
700
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCAmelCase ( _snake_case ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('''/''' ) lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase = torch.load(os.path.join(_snake_case , '''pytorch_model.bin''' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase , lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_snake_case ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F"""bertarized_{os.path.basename(_snake_case )}""" ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(_snake_case , os.path.join(_snake_case , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) UpperCAmelCase_ =parser.parse_args() main(args)
33
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def UpperCAmelCase ( _snake_case ): return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class __UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' @staticmethod def __snake_case ( UpperCAmelCase_ ): lowerCAmelCase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=__lowerCamelCase , default=__lowerCamelCase , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=__lowerCamelCase , help='''Name of the model to download''' ) download_parser.set_defaults(func=__lowerCamelCase ) def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = model lowerCAmelCase = cache lowerCAmelCase = force lowerCAmelCase = trust_remote_code def __snake_case ( self ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
701
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } UpperCAmelCase_ ={ """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } UpperCAmelCase_ ={ """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = set() lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase = char lowerCAmelCase = set(_snake_case ) return pairs class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =VOCAB_FILES_NAMES __a : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP __a : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , **UpperCAmelCase_ , ): super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = merges_file lowerCAmelCase = {} lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 lowerCAmelCase = 3 self.add_from_file(UpperCAmelCase_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase_ , encoding='''utf-8''' ) as merges_handle: lowerCAmelCase = merges_handle.read().split('''\n''' )[:-1] lowerCAmelCase = [tuple(merge.split()[:-1] ) for merge in merges] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __snake_case ( self ): return len(self.encoder ) def __snake_case ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self , UpperCAmelCase_ ): if token in self.cache: return self.cache[token] lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowerCAmelCase = get_pairs(UpperCAmelCase_ ) if not pairs: return token while True: lowerCAmelCase = min(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : self.bpe_ranks.get(UpperCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase = bigram lowerCAmelCase = [] lowerCAmelCase = 0 while i < len(UpperCAmelCase_ ): try: lowerCAmelCase = word.index(UpperCAmelCase_ , UpperCAmelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase = j if word[i] == first and i < len(UpperCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = new_word if len(UpperCAmelCase_ ) == 1: break else: lowerCAmelCase = get_pairs(UpperCAmelCase_ ) lowerCAmelCase = '''@@ '''.join(UpperCAmelCase_ ) lowerCAmelCase = word[:-4] lowerCAmelCase = word return word def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [] lowerCAmelCase = re.findall(r'''\S+\n?''' , UpperCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __snake_case ( self , UpperCAmelCase_ ): return self.encoder.get(UpperCAmelCase_ , self.encoder.get(self.unk_token ) ) def __snake_case ( self , UpperCAmelCase_ ): return self.decoder.get(UpperCAmelCase_ , self.unk_token ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = ''' '''.join(UpperCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.merges_file , UpperCAmelCase_ ) return out_vocab_file, out_merge_file def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(UpperCAmelCase_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCAmelCase = f.readlines() for lineTmp in lines: lowerCAmelCase = lineTmp.strip() lowerCAmelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowerCAmelCase = line[:idx] lowerCAmelCase = len(self.encoder )
33
0
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=7 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=99 , UpperCAmelCase_=32 , UpperCAmelCase_=2 , UpperCAmelCase_=4 , UpperCAmelCase_=37 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=5_12 , UpperCAmelCase_=16 , UpperCAmelCase_=2 , UpperCAmelCase_=0.02 , UpperCAmelCase_=3 , UpperCAmelCase_=4 , UpperCAmelCase_=None , UpperCAmelCase_=0 , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_input_mask lowerCAmelCase = use_token_type_ids lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = num_labels lowerCAmelCase = num_choices lowerCAmelCase = scope lowerCAmelCase = projection_dim def __snake_case ( self ): lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) lowerCAmelCase = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = TFDPRContextEncoder(config=_lowercase ) lowerCAmelCase = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase ) lowerCAmelCase = model(_lowercase , token_type_ids=_lowercase ) lowerCAmelCase = model(_lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = TFDPRQuestionEncoder(config=_lowercase ) lowerCAmelCase = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase ) lowerCAmelCase = model(_lowercase , token_type_ids=_lowercase ) lowerCAmelCase = model(_lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = TFDPRReader(config=_lowercase ) lowerCAmelCase = model(_lowercase , attention_mask=_lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = {'''input_ids''': input_ids} return config, inputs_dict @require_tf class __UpperCamelCase ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' __a : Tuple =( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) __a : Optional[int] ={"""feature-extraction""": TFDPRQuestionEncoder} if is_tf_available() else {} __a : Optional[int] =False __a : str =False __a : str =False __a : Any =False __a : str =False def __snake_case ( self ): lowerCAmelCase = TFDPRModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=_lowercase , hidden_size=37 ) def __snake_case ( self ): self.config_tester.run_common_tests() def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*_lowercase ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*_lowercase ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*_lowercase ) @slow def __snake_case ( self ): for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = TFDPRContextEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = TFDPRContextEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = TFDPRQuestionEncoder.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = TFDPRReader.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) @require_tf class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __snake_case ( self ): lowerCAmelCase = TFDPRQuestionEncoder.from_pretrained('''facebook/dpr-question_encoder-single-nq-base''' ) lowerCAmelCase = tf.constant( [[1_01, 75_92, 10_10, 20_03, 20_26, 38_99, 1_01_40, 10_29, 1_02]] ) # [CLS] hello, is my dog cute? [SEP] lowerCAmelCase = model(_lowercase )[0] # embedding shape = (1, 768) # compare the actual values for a slice. lowerCAmelCase = tf.constant( [ [ 0.03236253, 0.12753335, 0.16818509, 0.00279786, 0.3896933, 0.24264945, 0.2178971, -0.02335227, -0.08481959, -0.14324117, ] ] ) self.assertTrue(numpy.allclose(output[:, :10].numpy() , expected_slice.numpy() , atol=1E-4 ) )
702
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ =TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = self lowerCAmelCase = 0 class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # create a new set with x as its member lowerCAmelCase = DisjointSetTreeNode(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): # find the set x belongs to (with path-compression) lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # helper function for union operation if nodea.rank > nodea.rank: lowerCAmelCase = nodea else: lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # merge 2 disjoint sets self.link(self.find_set(UpperCAmelCase_ ) , self.find_set(UpperCAmelCase_ ) ) class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # add a node ONLY if its not present in the graph if node not in self.connections: lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # add an edge with the given weight self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) lowerCAmelCase = weight lowerCAmelCase = weight def __snake_case ( self ): lowerCAmelCase = [] lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCAmelCase_ : x[2] ) # creating the disjoint set lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(UpperCAmelCase_ ) # MST generation lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = edges[index] index += 1 lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) disjoint_set.union(UpperCAmelCase_ , UpperCAmelCase_ ) return graph
33
0
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( snake_case__ ): '''simple docstring''' __a : Optional[int] =CLIPConfig __a : List[Any] =["""CLIPEncoderLayer"""] def __init__( self , UpperCAmelCase_ ): super().__init__(lowercase_ ) lowerCAmelCase = CLIPVisionModelWithProjection(config.vision_config ) lowerCAmelCase = nn.Linear(config.vision_config.projection_dim , 1 ) lowerCAmelCase = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=0.5 , UpperCAmelCase_=0.5 ): lowerCAmelCase = self.vision_model(lowercase_ )[0] lowerCAmelCase = self.p_head(lowercase_ ) lowerCAmelCase = nsfw_detected.flatten() lowerCAmelCase = nsfw_detected > p_threshold lowerCAmelCase = nsfw_detected.tolist() if any(lowercase_ ): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''' ) for idx, nsfw_detected_ in enumerate(lowercase_ ): if nsfw_detected_: lowerCAmelCase = np.zeros(images[idx].shape ) lowerCAmelCase = self.w_head(lowercase_ ) lowerCAmelCase = watermark_detected.flatten() lowerCAmelCase = watermark_detected > w_threshold lowerCAmelCase = watermark_detected.tolist() if any(lowercase_ ): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''' ) for idx, watermark_detected_ in enumerate(lowercase_ ): if watermark_detected_: lowerCAmelCase = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
703
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations(_snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations_with_dp_array( _snake_case , _snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _snake_case ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =3 UpperCAmelCase_ =5 UpperCAmelCase_ =[1, 2, 5] print(combination_sum_iv(n, array, target))
33
0
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class __UpperCamelCase ( _UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = params lowerCAmelCase = np.array(A_ ) lowerCAmelCase = np.array([len(A_ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self , UpperCAmelCase_ ): return (self.token_ids[index], self.lengths[index]) def __len__( self ): return len(self.lengths ) def __snake_case ( self ): assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def __snake_case ( self ): lowerCAmelCase = self.params.max_model_input_size lowerCAmelCase = self.lengths > max_len logger.info(F"""Splitting {sum(A_ )} too long sequences.""" ) def divide_chunks(UpperCAmelCase_ , UpperCAmelCase_ ): return [l[i : i + n] for i in range(0 , len(A_ ) , A_ )] lowerCAmelCase = [] lowerCAmelCase = [] if self.params.mlm: lowerCAmelCase , lowerCAmelCase = self.params.special_tok_ids['''cls_token'''], self.params.special_tok_ids['''sep_token'''] else: lowerCAmelCase , lowerCAmelCase = self.params.special_tok_ids['''bos_token'''], self.params.special_tok_ids['''eos_token'''] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase = np.insert(A_ , 0 , A_ ) if sub_s[-1] != sep_id: lowerCAmelCase = np.insert(A_ , len(A_ ) , A_ ) assert len(A_ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(A_ ) new_tok_ids.extend(A_ ) new_lengths.extend([len(A_ ) for l in sub_seqs] ) lowerCAmelCase = np.array(A_ ) lowerCAmelCase = np.array(A_ ) def __snake_case ( self ): lowerCAmelCase = len(self ) lowerCAmelCase = self.lengths > 11 lowerCAmelCase = self.token_ids[indices] lowerCAmelCase = self.lengths[indices] lowerCAmelCase = len(self ) logger.info(F"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def __snake_case ( self ): if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase = self.params.special_tok_ids['''unk_token'''] lowerCAmelCase = len(self ) lowerCAmelCase = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase = (unk_occs / self.lengths) < 0.5 lowerCAmelCase = self.token_ids[indices] lowerCAmelCase = self.lengths[indices] lowerCAmelCase = len(self ) logger.info(F"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def __snake_case ( self ): if not self.params.is_master: return logger.info(F"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [t[0] for t in batch] lowerCAmelCase = [t[1] for t in batch] assert len(A_ ) == len(A_ ) # Max for paddings lowerCAmelCase = max(A_ ) # Pad token ids if self.params.mlm: lowerCAmelCase = self.params.special_tok_ids['''pad_token'''] else: lowerCAmelCase = self.params.special_tok_ids['''unk_token'''] lowerCAmelCase = [list(t.astype(A_ ) ) + [pad_idx] * (max_seq_len_ - len(A_ )) for t in token_ids] assert len(tk_ ) == len(A_ ) assert all(len(A_ ) == max_seq_len_ for t in tk_ ) lowerCAmelCase = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase = torch.tensor(A_ ) # (bs) return tk_t, lg_t
704
import torch from diffusers import StableDiffusionPipeline UpperCAmelCase_ ="""path-to-your-trained-model""" UpperCAmelCase_ =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") UpperCAmelCase_ ="""A photo of sks dog in a bucket""" UpperCAmelCase_ =pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
33
0
'''simple docstring''' def UpperCAmelCase ( _snake_case ): lowerCAmelCase = len(_lowerCamelCase ) lowerCAmelCase = sum(_lowerCamelCase ) lowerCAmelCase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): lowerCAmelCase = True for i in range(1 , s + 1 ): lowerCAmelCase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): lowerCAmelCase = dp[i][j - 1] if arr[i - 1] <= j: lowerCAmelCase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: lowerCAmelCase = s - 2 * j break return diff
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def UpperCAmelCase ( _snake_case ): lowerCAmelCase = SwinConfig(image_size=192 ) if "base" in model_name: lowerCAmelCase = 6 lowerCAmelCase = 128 lowerCAmelCase = (2, 2, 18, 2) lowerCAmelCase = (4, 8, 16, 32) elif "large" in model_name: lowerCAmelCase = 12 lowerCAmelCase = 192 lowerCAmelCase = (2, 2, 18, 2) lowerCAmelCase = (6, 12, 24, 48) else: raise ValueError('''Model not supported, only supports base and large variants''' ) lowerCAmelCase = window_size lowerCAmelCase = embed_dim lowerCAmelCase = depths lowerCAmelCase = num_heads return config def UpperCAmelCase ( _snake_case ): if "encoder.mask_token" in name: lowerCAmelCase = name.replace('''encoder.mask_token''' , '''embeddings.mask_token''' ) if "encoder.patch_embed.proj" in name: lowerCAmelCase = name.replace('''encoder.patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "encoder.patch_embed.norm" in name: lowerCAmelCase = name.replace('''encoder.patch_embed.norm''' , '''embeddings.norm''' ) if "attn.proj" in name: lowerCAmelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowerCAmelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowerCAmelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowerCAmelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowerCAmelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowerCAmelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) if name == "encoder.norm.weight": lowerCAmelCase = '''layernorm.weight''' if name == "encoder.norm.bias": lowerCAmelCase = '''layernorm.bias''' if "decoder" in name: pass else: lowerCAmelCase = '''swin.''' + name return name def UpperCAmelCase ( _snake_case , _snake_case ): for key in orig_state_dict.copy().keys(): lowerCAmelCase = orig_state_dict.pop(_SCREAMING_SNAKE_CASE ) if "attn_mask" in key: pass elif "qkv" in key: lowerCAmelCase = key.split('''.''' ) lowerCAmelCase = int(key_split[2] ) lowerCAmelCase = int(key_split[4] ) lowerCAmelCase = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCAmelCase = val[:dim, :] lowerCAmelCase = val[ dim : dim * 2, : ] lowerCAmelCase = val[-dim:, :] else: lowerCAmelCase = val[ :dim ] lowerCAmelCase = val[ dim : dim * 2 ] lowerCAmelCase = val[ -dim: ] else: lowerCAmelCase = val return orig_state_dict def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase = torch.load(_SCREAMING_SNAKE_CASE , map_location='''cpu''' )['''model'''] lowerCAmelCase = get_swin_config(_SCREAMING_SNAKE_CASE ) lowerCAmelCase = SwinForMaskedImageModeling(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase = convert_state_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) lowerCAmelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase = ViTImageProcessor(size={'''height''': 192, '''width''': 192} ) lowerCAmelCase = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) lowerCAmelCase = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) with torch.no_grad(): lowerCAmelCase = model(**_SCREAMING_SNAKE_CASE ).logits print(outputs.keys() ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and image processor for {model_name} to hub""" ) model.push_to_hub(F"""microsoft/{model_name}""" ) image_processor.push_to_hub(F"""microsoft/{model_name}""" ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""swin-base-simmim-window6-192""", type=str, choices=["""swin-base-simmim-window6-192""", """swin-large-simmim-window12-192"""], help="""Name of the Swin SimMIM model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth""", type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) UpperCAmelCase_ =parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
706
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = 8 # DPR tok lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __snake_case ( self ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __snake_case ( self ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: lowerCAmelCase = os.path.join(self.tmpdirname , '''dataset''' ) lowerCAmelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCAmelCase_ ) , ) return retriever def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) lowerCAmelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) lowerCAmelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(UpperCAmelCase_ , open(UpperCAmelCase_ , '''wb''' ) ) lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = self.get_dummy_dataset() retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_legacy_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): import torch lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) lowerCAmelCase = retriever( UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ , return_tensors='''pt''' , ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): lowerCAmelCase = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) retriever.set_ctx_encoder_tokenizer(UpperCAmelCase_ ) lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) self.assertEqual( len(UpperCAmelCase_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCAmelCase_ ) # check for doc token related keys in dictionary.
33
0
import requests UpperCAmelCase_ ="""""" # <-- Put your OpenWeatherMap appid here! UpperCAmelCase_ ="""https://api.openweathermap.org/data/2.5/""" def UpperCAmelCase ( _snake_case = "Chicago" , _snake_case = APPID ): return requests.get(URL_BASE + '''weather''' , params=locals() ).json() def UpperCAmelCase ( _snake_case = "Kolkata, India" , _snake_case = APPID ): return requests.get(URL_BASE + '''forecast''' , params=locals() ).json() def UpperCAmelCase ( _snake_case = 55.68 , _snake_case = 12.57 , _snake_case = APPID ): return requests.get(URL_BASE + '''onecall''' , params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: UpperCAmelCase_ =input("""Enter a location:""").strip() if location: pprint(current_weather(location)) else: break
707
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """google/switch-base-8""": """https://huggingface.co/google/switch-base-8/blob/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ="""switch_transformers""" __a : Union[str, Any] =["""past_key_values"""] __a : Dict ={"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , UpperCAmelCase_=3_21_28 , UpperCAmelCase_=7_68 , UpperCAmelCase_=64 , UpperCAmelCase_=20_48 , UpperCAmelCase_=64 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=8 , UpperCAmelCase_=False , UpperCAmelCase_=0.01 , UpperCAmelCase_="float32" , UpperCAmelCase_=False , UpperCAmelCase_=32 , UpperCAmelCase_=1_28 , UpperCAmelCase_=0.1 , UpperCAmelCase_=1E-6 , UpperCAmelCase_=0.001 , UpperCAmelCase_=0.001 , UpperCAmelCase_=1.0 , UpperCAmelCase_="relu" , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=0 , UpperCAmelCase_=1 , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = d_kv lowerCAmelCase = d_ff lowerCAmelCase = num_sparse_encoder_layers lowerCAmelCase = num_layers lowerCAmelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: lowerCAmelCase = self.num_layers // self.num_sparse_encoder_layers else: lowerCAmelCase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: lowerCAmelCase = self.num_decoder_layers // self.num_sparse_decoder_layers else: lowerCAmelCase = self.num_decoder_layers # HACK: this will create 0 sparse layers lowerCAmelCase = num_heads lowerCAmelCase = num_experts lowerCAmelCase = expert_capacity lowerCAmelCase = router_bias lowerCAmelCase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) lowerCAmelCase = router_dtype lowerCAmelCase = router_ignore_padding_tokens lowerCAmelCase = relative_attention_num_buckets lowerCAmelCase = relative_attention_max_distance lowerCAmelCase = dropout_rate lowerCAmelCase = layer_norm_epsilon lowerCAmelCase = initializer_factor lowerCAmelCase = feed_forward_proj lowerCAmelCase = use_cache lowerCAmelCase = add_router_probs lowerCAmelCase = router_z_loss_coef lowerCAmelCase = router_aux_loss_coef lowerCAmelCase = self.feed_forward_proj.split('''-''' ) lowerCAmelCase = act_info[-1] lowerCAmelCase = act_info[0] == '''gated''' if len(UpperCAmelCase_ ) > 1 and act_info[0] != "gated" or len(UpperCAmelCase_ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCAmelCase = '''gelu_new''' super().__init__( pad_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ , )
33
0
UpperCAmelCase_ =0 # The first color of the flag. UpperCAmelCase_ =1 # The second color of the flag. UpperCAmelCase_ =2 # The third color of the flag. UpperCAmelCase_ =(red, white, blue) def UpperCAmelCase ( _snake_case ): if not sequence: return [] if len(lowerCAmelCase__ ) == 1: return list(lowerCAmelCase__ ) lowerCAmelCase = 0 lowerCAmelCase = len(lowerCAmelCase__ ) - 1 lowerCAmelCase = 0 while mid <= high: if sequence[mid] == colors[0]: lowerCAmelCase = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: lowerCAmelCase = sequence[high], sequence[mid] high -= 1 else: lowerCAmelCase = F"""The elements inside the sequence must contains only {colors} values""" raise ValueError(lowerCAmelCase__ ) return sequence if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =input("""Enter numbers separated by commas:\n""").strip() UpperCAmelCase_ =[int(item.strip()) for item in user_input.split(""",""")] print(F'''{dutch_national_flag_sort(unsorted)}''')
708
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def UpperCAmelCase ( _snake_case ): lowerCAmelCase , lowerCAmelCase = analyze_text(_snake_case ) lowerCAmelCase = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase = sum(single_char_strings.values() ) # one length string lowerCAmelCase = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase = single_char_strings[ch] lowerCAmelCase = my_str / all_sum my_fir_sum += prob * math.loga(_snake_case ) # entropy formula. # print entropy print(F"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase = sum(two_char_strings.values() ) lowerCAmelCase = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase = cha + cha if sequence in two_char_strings: lowerCAmelCase = two_char_strings[sequence] lowerCAmelCase = int(_snake_case ) / all_sum my_sec_sum += prob * math.loga(_snake_case ) # print second entropy print(F"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(F"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = Counter() # type: ignore lowerCAmelCase = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(_snake_case ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def UpperCAmelCase ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
33
0
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ =logging.get_logger(__name__) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: lowerCAmelCase = 128 elif "12-12" in model_name: lowerCAmelCase = 12 lowerCAmelCase = 12 elif "14-14" in model_name: lowerCAmelCase = 14 lowerCAmelCase = 14 elif "16-16" in model_name: lowerCAmelCase = 16 lowerCAmelCase = 16 else: raise ValueError('''Model not supported''' ) lowerCAmelCase = '''huggingface/label-files''' if "speech-commands" in model_name: lowerCAmelCase = 35 lowerCAmelCase = '''speech-commands-v2-id2label.json''' else: lowerCAmelCase = 527 lowerCAmelCase = '''audioset-id2label.json''' lowerCAmelCase = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase = {int(_snake_case ): v for k, v in idalabel.items()} lowerCAmelCase = idalabel lowerCAmelCase = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase ( _snake_case ): if "module.v" in name: lowerCAmelCase = name.replace('''module.v''' , '''audio_spectrogram_transformer''' ) if "cls_token" in name: lowerCAmelCase = name.replace('''cls_token''' , '''embeddings.cls_token''' ) if "dist_token" in name: lowerCAmelCase = name.replace('''dist_token''' , '''embeddings.distillation_token''' ) if "pos_embed" in name: lowerCAmelCase = name.replace('''pos_embed''' , '''embeddings.position_embeddings''' ) if "patch_embed.proj" in name: lowerCAmelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) # transformer blocks if "blocks" in name: lowerCAmelCase = name.replace('''blocks''' , '''encoder.layer''' ) if "attn.proj" in name: lowerCAmelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowerCAmelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowerCAmelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowerCAmelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowerCAmelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowerCAmelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) # final layernorm if "audio_spectrogram_transformer.norm" in name: lowerCAmelCase = name.replace('''audio_spectrogram_transformer.norm''' , '''audio_spectrogram_transformer.layernorm''' ) # classifier head if "module.mlp_head.0" in name: lowerCAmelCase = name.replace('''module.mlp_head.0''' , '''classifier.layernorm''' ) if "module.mlp_head.1" in name: lowerCAmelCase = name.replace('''module.mlp_head.1''' , '''classifier.dense''' ) return name def UpperCAmelCase ( _snake_case , _snake_case ): for key in orig_state_dict.copy().keys(): lowerCAmelCase = orig_state_dict.pop(_snake_case ) if "qkv" in key: lowerCAmelCase = key.split('''.''' ) lowerCAmelCase = int(key_split[3] ) lowerCAmelCase = config.hidden_size if "weight" in key: lowerCAmelCase = val[:dim, :] lowerCAmelCase = val[dim : dim * 2, :] lowerCAmelCase = val[-dim:, :] else: lowerCAmelCase = val[:dim] lowerCAmelCase = val[dim : dim * 2] lowerCAmelCase = val[-dim:] else: lowerCAmelCase = val return orig_state_dict def UpperCAmelCase ( _snake_case ): lowerCAmelCase = [ '''module.v.head.weight''', '''module.v.head.bias''', '''module.v.head_dist.weight''', '''module.v.head_dist.bias''', ] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) @torch.no_grad() def UpperCAmelCase ( _snake_case , _snake_case , _snake_case=False ): lowerCAmelCase = get_audio_spectrogram_transformer_config(_snake_case ) lowerCAmelCase = { '''ast-finetuned-audioset-10-10-0.4593''': ( '''https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.450''': ( '''https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.448''': ( '''https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.448-v2''': ( '''https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1''' ), '''ast-finetuned-audioset-12-12-0.447''': ( '''https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1''' ), '''ast-finetuned-audioset-14-14-0.443''': ( '''https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1''' ), '''ast-finetuned-audioset-16-16-0.442''': ( '''https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1''' ), '''ast-finetuned-speech-commands-v2''': ( '''https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1''' ), } # load original state_dict lowerCAmelCase = model_name_to_url[model_name] lowerCAmelCase = torch.hub.load_state_dict_from_url(_snake_case , map_location='''cpu''' ) # remove some keys remove_keys(_snake_case ) # rename some keys lowerCAmelCase = convert_state_dict(_snake_case , _snake_case ) # load 🤗 model lowerCAmelCase = ASTForAudioClassification(_snake_case ) model.eval() model.load_state_dict(_snake_case ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 lowerCAmelCase = -4.2677393 if '''speech-commands''' not in model_name else -6.845978 lowerCAmelCase = 4.5689974 if '''speech-commands''' not in model_name else 5.5654526 lowerCAmelCase = 1024 if '''speech-commands''' not in model_name else 128 lowerCAmelCase = ASTFeatureExtractor(mean=_snake_case , std=_snake_case , max_length=_snake_case ) if "speech-commands" in model_name: lowerCAmelCase = load_dataset('''speech_commands''' , '''v0.02''' , split='''validation''' ) lowerCAmelCase = dataset[0]['''audio''']['''array'''] else: lowerCAmelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' , ) lowerCAmelCase , lowerCAmelCase = torchaudio.load(_snake_case ) lowerCAmelCase = waveform.squeeze().numpy() lowerCAmelCase = feature_extractor(_snake_case , sampling_rate=16000 , return_tensors='''pt''' ) # forward pass lowerCAmelCase = model(**_snake_case ) lowerCAmelCase = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": lowerCAmelCase = torch.tensor([-0.8760, -7.0042, -8.6602] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": lowerCAmelCase = torch.tensor([-1.1986, -7.0903, -8.2718] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": lowerCAmelCase = torch.tensor([-2.6128, -8.0080, -9.4344] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": lowerCAmelCase = torch.tensor([-1.5080, -7.4534, -8.8917] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": lowerCAmelCase = torch.tensor([-0.5050, -6.5833, -8.0843] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": lowerCAmelCase = torch.tensor([-0.3826, -7.0336, -8.2413] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": lowerCAmelCase = torch.tensor([-1.2113, -6.9101, -8.3470] ) elif model_name == "ast-finetuned-speech-commands-v2": lowerCAmelCase = torch.tensor([6.1589, -8.0566, -8.7984] ) else: raise ValueError('''Unknown model name''' ) if not torch.allclose(logits[0, :3] , _snake_case , atol=1E-4 ): raise ValueError('''Logits don\'t match''' ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_snake_case ) print(F"""Saving feature extractor to {pytorch_dump_folder_path}""" ) feature_extractor.save_pretrained(_snake_case ) if push_to_hub: print('''Pushing model and feature extractor to the hub...''' ) model.push_to_hub(F"""MIT/{model_name}""" ) feature_extractor.push_to_hub(F"""MIT/{model_name}""" ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""ast-finetuned-audioset-10-10-0.4593""", type=str, help="""Name of the Audio Spectrogram Transformer model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) UpperCAmelCase_ =parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
709
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Tuple =IFInpaintingSuperResolutionPipeline __a : Dict =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __a : int =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) __a : Union[str, Any] =PipelineTesterMixin.required_optional_params - {"""latents"""} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
33
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: UpperCAmelCase_ =None UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ="""▁""" UpperCAmelCase_ ={"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} UpperCAmelCase_ ={ """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } UpperCAmelCase_ ={ """google/pegasus-xsum""": 512, } class __UpperCamelCase ( lowercase_ ): '''simple docstring''' __a : Optional[int] =VOCAB_FILES_NAMES __a : int =PRETRAINED_VOCAB_FILES_MAP __a : Optional[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __a : List[str] =PegasusTokenizer __a : int =['''input_ids''', '''attention_mask'''] def __init__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_="<pad>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<mask_2>" , UpperCAmelCase_="<mask_1>" , UpperCAmelCase_=None , UpperCAmelCase_=1_03 , **UpperCAmelCase_ , ): lowerCAmelCase = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError( F"""additional_special_tokens should be of type {type(UpperCamelCase__ )}, but is""" F""" {type(UpperCamelCase__ )}""" ) lowerCAmelCase = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"""<unk_{i}>""" for i in range(len(UpperCamelCase__ ) , self.offset - 1 ) ] if len(set(UpperCamelCase__ ) ) != len(UpperCamelCase__ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' F""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) lowerCAmelCase = additional_special_tokens_extended else: lowerCAmelCase = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , pad_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , mask_token_sent=UpperCamelCase__ , offset=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCAmelCase = vocab_file lowerCAmelCase = False if not self.vocab_file else True def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( '''There should be 3 special tokens: mask_token, pad_token, and eos_token +''' F""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return self._special_token_mask(UpperCamelCase__ ) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
710
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ ={ """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =StableDiffusionDiffEditPipeline __a : List[str] =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} __a : Union[str, Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} __a : int =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __a : str =frozenset([] ) def __snake_case ( self ): torch.manual_seed(0 ) lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCAmelCase_ , ) lowerCAmelCase = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCAmelCase_ , set_alpha_to_one=lowerCAmelCase_ , ) lowerCAmelCase = DDIMInverseScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCAmelCase_ , set_alpha_to_zero=lowerCAmelCase_ , ) torch.manual_seed(0 ) lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=5_12 , ) lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) lowerCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase = { '''unet''': unet, '''scheduler''': scheduler, '''inverse_scheduler''': inverse_scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 16, 16) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) if str(lowerCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''a dog and a newt''', '''mask_image''': mask, '''image_latents''': latents, '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('''RGB''' ) if str(lowerCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) lowerCAmelCase = { '''image''': image, '''source_prompt''': '''a cat and a frog''', '''target_prompt''': '''a dog and a newt''', '''generator''': generator, '''num_inference_steps''': 2, '''num_maps_per_mask''': 2, '''mask_encode_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('''RGB''' ) if str(lowerCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) lowerCAmelCase = { '''image''': image, '''prompt''': '''a cat and a frog''', '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''decode_latents''': True, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): if not hasattr(self.pipeline_class , '''_optional_components''' ): return lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = self.pipeline_class(**lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCAmelCase_ ) lowerCAmelCase = self.pipeline_class.from_pretrained(lowerCAmelCase_ ) pipe_loaded.to(lowerCAmelCase_ ) pipe_loaded.set_progress_bar_config(disable=lowerCAmelCase_ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCAmelCase_ , lowerCAmelCase_ ) is None , F"""`{optional_component}` did not stay set to None after loading.""" , ) lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) lowerCAmelCase = pipe_loaded(**lowerCAmelCase_ )[0] lowerCAmelCase = np.abs(output - output_loaded ).max() self.assertLess(lowerCAmelCase_ , 1E-4 ) def __snake_case ( self ): lowerCAmelCase = '''cpu''' lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = self.pipeline_class(**lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) lowerCAmelCase = self.get_dummy_mask_inputs(lowerCAmelCase_ ) lowerCAmelCase = pipe.generate_mask(**lowerCAmelCase_ ) lowerCAmelCase = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) lowerCAmelCase = np.array([0] * 9 ) lowerCAmelCase = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase_ , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def __snake_case ( self ): lowerCAmelCase = '''cpu''' lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = self.pipeline_class(**lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) lowerCAmelCase = self.get_dummy_inversion_inputs(lowerCAmelCase_ ) lowerCAmelCase = pipe.invert(**lowerCAmelCase_ ).images lowerCAmelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) lowerCAmelCase = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.51050, 0.5015, 0.4407, 0.4799] , ) lowerCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase_ , 1E-3 ) def __snake_case ( self ): super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def __snake_case ( self ): lowerCAmelCase = '''cpu''' lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = {'''beta_start''': 0.00085, '''beta_end''': 0.012, '''beta_schedule''': '''scaled_linear'''} lowerCAmelCase = DPMSolverMultistepScheduler(**lowerCAmelCase_ ) lowerCAmelCase = DPMSolverMultistepInverseScheduler(**lowerCAmelCase_ ) lowerCAmelCase = self.pipeline_class(**lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) lowerCAmelCase = self.get_dummy_inversion_inputs(lowerCAmelCase_ ) lowerCAmelCase = pipe.invert(**lowerCAmelCase_ ).images lowerCAmelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) lowerCAmelCase = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.51050, 0.5015, 0.4407, 0.4799] , ) lowerCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase_ , 1E-3 ) @require_torch_gpu @slow class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __snake_case ( cls ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png''' ) lowerCAmelCase = raw_image.convert('''RGB''' ).resize((7_68, 7_68) ) lowerCAmelCase = raw_image def __snake_case ( self ): lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) lowerCAmelCase = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) lowerCAmelCase = '''a bowl of fruit''' lowerCAmelCase = '''a bowl of pears''' lowerCAmelCase = pipe.generate_mask( image=self.raw_image , source_prompt=lowerCAmelCase_ , target_prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , ) lowerCAmelCase = pipe.invert( prompt=lowerCAmelCase_ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCAmelCase_ ).latents lowerCAmelCase = pipe( prompt=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , image_latents=lowerCAmelCase_ , generator=lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ , inpaint_strength=0.7 , output_type='''numpy''' , ).images[0] lowerCAmelCase = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5E-1 def __snake_case ( self ): lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) lowerCAmelCase = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) lowerCAmelCase = '''a bowl of fruit''' lowerCAmelCase = '''a bowl of pears''' lowerCAmelCase = pipe.generate_mask( image=self.raw_image , source_prompt=lowerCAmelCase_ , target_prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , ) lowerCAmelCase = pipe.invert( prompt=lowerCAmelCase_ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCAmelCase_ , num_inference_steps=25 , ).latents lowerCAmelCase = pipe( prompt=lowerCAmelCase_ , mask_image=lowerCAmelCase_ , image_latents=lowerCAmelCase_ , generator=lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ , inpaint_strength=0.7 , num_inference_steps=25 , output_type='''numpy''' , ).images[0] lowerCAmelCase = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5E-1
711
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline UpperCAmelCase_ =datasets.utils.logging.get_logger(__name__) @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __a : Optional[datasets.Features] =None __a : str ="utf-8" __a : Optional[str] =None __a : Optional[str] =None __a : bool =True # deprecated __a : Optional[int] =None # deprecated __a : int =1_0 << 2_0 # 10MB __a : Optional[bool] =None class __UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __a : str =JsonConfig def __snake_case ( self ): if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowerCAmelCase = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def __snake_case ( self , UpperCAmelCase_ ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase_ , (str, list, tuple) ): lowerCAmelCase = data_files if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase_ , gen_kwargs={'''files''': files} ) ) return splits def __snake_case ( self , UpperCAmelCase_ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCAmelCase = self.config.features.arrow_schema.field(UpperCAmelCase_ ).type lowerCAmelCase = pa_table.append_column(UpperCAmelCase_ , pa.array([None] * len(UpperCAmelCase_ ) , type=UpperCAmelCase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase = table_cast(UpperCAmelCase_ , self.config.features.arrow_schema ) return pa_table def __snake_case ( self , UpperCAmelCase_ ): for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) # We keep only the field we are interested in lowerCAmelCase = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(UpperCAmelCase_ , (list, tuple) ): lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} else: lowerCAmelCase = dataset lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) yield file_idx, self._cast_table(UpperCAmelCase_ ) # If the file has one json object per line else: with open(UpperCAmelCase_ , '''rb''' ) as f: lowerCAmelCase = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCAmelCase = max(self.config.chunksize // 32 , 16 << 10 ) lowerCAmelCase = ( self.config.encoding_errors if self.config.encoding_errors is not None else '''strict''' ) while True: lowerCAmelCase = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(UpperCAmelCase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCAmelCase = batch.decode(self.config.encoding , errors=UpperCAmelCase_ ).encode('''utf-8''' ) try: while True: try: lowerCAmelCase = paj.read_json( io.BytesIO(UpperCAmelCase_ ) , read_options=paj.ReadOptions(block_size=UpperCAmelCase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(UpperCAmelCase_ , pa.ArrowInvalid ) and "straddling" not in str(UpperCAmelCase_ ) or block_size > len(UpperCAmelCase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(UpperCAmelCase_ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # list is the only sequence type supported in JSON try: lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(UpperCAmelCase_ ) break else: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase_ ) batch_idx += 1
33
0
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase_ =get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class __UpperCamelCase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __a : Optional[Any] =XLMRobertaTokenizer __a : Any =XLMRobertaTokenizerFast __a : Optional[int] =True __a : str =True def __snake_case ( self ): super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase = XLMRobertaTokenizer(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = '''<pad>''' lowerCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def __snake_case ( self ): lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(_UpperCAmelCase ) , 10_02 ) def __snake_case ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def __snake_case ( self ): lowerCAmelCase = XLMRobertaTokenizer(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) lowerCAmelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) lowerCAmelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowerCAmelCase = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) self.assertListEqual( _UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCAmelCase = tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual( _UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def __snake_case ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCAmelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) lowerCAmelCase = self.tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.save_pretrained(_UpperCAmelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(_UpperCAmelCase , _UpperCAmelCase ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.from_pretrained(_UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_UpperCAmelCase , _UpperCAmelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_UpperCAmelCase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(_UpperCAmelCase , legacy_format=_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.save_pretrained(_UpperCAmelCase ) # Checks it save with the same files self.assertSequenceEqual(_UpperCAmelCase , _UpperCAmelCase ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.from_pretrained(_UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_UpperCAmelCase , _UpperCAmelCase ) ) shutil.rmtree(_UpperCAmelCase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = tokenizer_r.save_pretrained(_UpperCAmelCase , legacy_format=_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.save_pretrained(_UpperCAmelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase = tokenizer_r.from_pretrained(_UpperCAmelCase ) lowerCAmelCase = tokenizer_p.from_pretrained(_UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_UpperCAmelCase , _UpperCAmelCase ) ) shutil.rmtree(_UpperCAmelCase ) @cached_property def __snake_case ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def __snake_case ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(_UpperCAmelCase , f.name ) lowerCAmelCase = XLMRobertaTokenizer(f.name , keep_accents=_UpperCAmelCase ) lowerCAmelCase = pickle.dumps(_UpperCAmelCase ) pickle.loads(_UpperCAmelCase ) def __snake_case ( self ): if not self.test_rust_tokenizer: return lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = '''I was born in 92000, and this is falsé.''' lowerCAmelCase = tokenizer.tokenize(_UpperCAmelCase ) lowerCAmelCase = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowerCAmelCase = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) lowerCAmelCase = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowerCAmelCase = self.get_rust_tokenizer() lowerCAmelCase = tokenizer.encode(_UpperCAmelCase ) lowerCAmelCase = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @slow def __snake_case ( self ): lowerCAmelCase = '''Hello World!''' lowerCAmelCase = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_UpperCAmelCase , self.big_tokenizer.encode(_UpperCAmelCase ) ) @slow def __snake_case ( self ): lowerCAmelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) lowerCAmelCase = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_UpperCAmelCase , self.big_tokenizer.encode(_UpperCAmelCase ) ) @slow def __snake_case ( self ): # fmt: off lowerCAmelCase = {'''input_ids''': [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
712
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] ="""maskformer-swin""" __a : Optional[int] ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , UpperCAmelCase_=2_24 , UpperCAmelCase_=4 , UpperCAmelCase_=3 , UpperCAmelCase_=96 , UpperCAmelCase_=[2, 2, 6, 2] , UpperCAmelCase_=[3, 6, 12, 24] , UpperCAmelCase_=7 , UpperCAmelCase_=4.0 , UpperCAmelCase_=True , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.1 , UpperCAmelCase_="gelu" , UpperCAmelCase_=False , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-5 , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = embed_dim lowerCAmelCase = depths lowerCAmelCase = len(UpperCAmelCase_ ) lowerCAmelCase = num_heads lowerCAmelCase = window_size lowerCAmelCase = mlp_ratio lowerCAmelCase = qkv_bias lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = drop_path_rate lowerCAmelCase = hidden_act lowerCAmelCase = use_absolute_embeddings lowerCAmelCase = layer_norm_eps lowerCAmelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase = int(embed_dim * 2 ** (len(UpperCAmelCase_ ) - 1) ) lowerCAmelCase = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase_ ) + 1 )] lowerCAmelCase , lowerCAmelCase = get_aligned_output_features_output_indices( out_features=UpperCAmelCase_ , out_indices=UpperCAmelCase_ , stage_names=self.stage_names )
33
0
def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = '''''' for i in table: res += inp[i - 1] return res def UpperCAmelCase ( _snake_case ): return data[1:] + data[0] def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = '''''' for i in range(len(UpperCAmelCase__ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = int('''0b''' + data[0] + data[-1] , 2 ) lowerCAmelCase = int('''0b''' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase = message[:4] lowerCAmelCase = message[4:] lowerCAmelCase = apply_table(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = xor(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = apply_sbox(UpperCAmelCase__ , temp[:4] ) # noqa: E741 lowerCAmelCase = apply_sbox(UpperCAmelCase__ , temp[4:] ) lowerCAmelCase = '''0''' * (2 - len(UpperCAmelCase__ )) + l # noqa: E741 lowerCAmelCase = '''0''' * (2 - len(UpperCAmelCase__ )) + r lowerCAmelCase = apply_table(l + r , UpperCAmelCase__ ) lowerCAmelCase = xor(UpperCAmelCase__ , UpperCAmelCase__ ) return temp + right if __name__ == "__main__": UpperCAmelCase_ =input("""Enter 10 bit key: """) UpperCAmelCase_ =input("""Enter 8 bit message: """) UpperCAmelCase_ =[6, 3, 7, 4, 8, 5, 10, 9] UpperCAmelCase_ =[3, 5, 2, 7, 4, 10, 1, 9, 8, 6] UpperCAmelCase_ =[2, 4, 3, 1] UpperCAmelCase_ =[2, 6, 3, 1, 4, 8, 5, 7] UpperCAmelCase_ =[4, 1, 3, 5, 7, 2, 8, 6] UpperCAmelCase_ =[4, 1, 2, 3, 2, 3, 4, 1] UpperCAmelCase_ =[[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] UpperCAmelCase_ =[[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation UpperCAmelCase_ =apply_table(key, paa_table) UpperCAmelCase_ =temp[:5] UpperCAmelCase_ =temp[5:] UpperCAmelCase_ =left_shift(left) UpperCAmelCase_ =left_shift(right) UpperCAmelCase_ =apply_table(left + right, pa_table) UpperCAmelCase_ =left_shift(left) UpperCAmelCase_ =left_shift(right) UpperCAmelCase_ =left_shift(left) UpperCAmelCase_ =left_shift(right) UpperCAmelCase_ =apply_table(left + right, pa_table) # encryption UpperCAmelCase_ =apply_table(message, IP) UpperCAmelCase_ =function(expansion, sa, sa, keya, temp) UpperCAmelCase_ =temp[4:] + temp[:4] UpperCAmelCase_ =function(expansion, sa, sa, keya, temp) UpperCAmelCase_ =apply_table(temp, IP_inv) print("""Cipher text is:""", CT) # decryption UpperCAmelCase_ =apply_table(CT, IP) UpperCAmelCase_ =function(expansion, sa, sa, keya, temp) UpperCAmelCase_ =temp[4:] + temp[:4] UpperCAmelCase_ =function(expansion, sa, sa, keya, temp) UpperCAmelCase_ =apply_table(temp, IP_inv) print("""Plain text after decypting is:""", PT)
713
from collections.abc import Sequence def UpperCAmelCase ( _snake_case , _snake_case = False ): if not arr: return 0 lowerCAmelCase = 0 if allow_empty_subarrays else float('''-inf''' ) lowerCAmelCase = 0.0 for num in arr: lowerCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) lowerCAmelCase = max(_snake_case , _snake_case ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase_ =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'''{max_subarray_sum(nums) = }''')
33
0
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=64 , UpperCAmelCase_=2 , UpperCAmelCase_=3 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=32 , UpperCAmelCase_=5 , UpperCAmelCase_=4 , UpperCAmelCase_=37 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=10 , UpperCAmelCase_=0.02 , UpperCAmelCase_=[1, 16, 4, 4] , UpperCAmelCase_=None , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = scope lowerCAmelCase = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size lowerCAmelCase = (self.image_size // 32) ** 2 lowerCAmelCase = num_patches + 1 def __snake_case ( self ): lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = self.get_config() return config, pixel_values, labels def __snake_case ( self ): lowerCAmelCase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [4, 8, 16, 32], '''num_groups''': 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=__UpperCamelCase , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = ViTHybridModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowerCAmelCase = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = self.type_sequence_label_size lowerCAmelCase = ViTHybridForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() lowerCAmelCase = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = config_and_inputs lowerCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Optional[Any] =(ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __a : Dict =( {"""feature-extraction""": ViTHybridModel, """image-classification""": ViTHybridForImageClassification} if is_torch_available() else {} ) __a : Optional[int] =False __a : Optional[int] =False __a : Optional[int] =False def __snake_case ( self ): lowerCAmelCase = ViTHybridModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __snake_case ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def __snake_case ( self ): pass def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(__UpperCamelCase ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: lowerCAmelCase = model_class(config=__UpperCamelCase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": lowerCAmelCase = [F"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @slow def __snake_case ( self ): for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = ViTHybridModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def UpperCAmelCase ( ): lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __snake_case ( self ): return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self ): lowerCAmelCase = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( __UpperCamelCase ) lowerCAmelCase = self.default_image_processor lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=__UpperCamelCase , return_tensors='''pt''' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**__UpperCamelCase ) # verify the logits lowerCAmelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) lowerCAmelCase = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow @require_accelerate def __snake_case ( self ): lowerCAmelCase = ViTHybridImageProcessor.from_pretrained('''google/vit-hybrid-base-bit-384''' ) lowerCAmelCase = ViTHybridForImageClassification.from_pretrained('''google/vit-hybrid-base-bit-384''' , device_map='''auto''' ) lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=__UpperCamelCase , return_tensors='''pt''' ) lowerCAmelCase = model(**__UpperCamelCase ) lowerCAmelCase = outputs.logits # model predicts one of the 1000 ImageNet classes lowerCAmelCase = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , '''tabby, tabby cat''' )
714
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =BertJapaneseTokenizer __a : Optional[int] =False __a : int =True def __snake_case ( self ): super().setUp() lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(do_lower_case=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer( do_lower_case=UpperCAmelCase_ , normalize_text=UpperCAmelCase_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(normalize_text=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(do_lower_case=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(normalize_text=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(trim_whitespace=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(do_lower_case=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(normalize_text=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(trim_whitespace=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = WordpieceTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) lowerCAmelCase = tokenizer.subword_tokenizer lowerCAmelCase = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) lowerCAmelCase = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =BertJapaneseTokenizer __a : Optional[int] =False def __snake_case ( self ): super().setUp() lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , **UpperCAmelCase_ ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( UpperCAmelCase_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = CharacterTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) lowerCAmelCase = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
33
0
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def UpperCAmelCase ( _snake_case ): '''simple docstring''' def is_in_circle(_snake_case , _snake_case ) -> bool: lowerCAmelCase = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle lowerCAmelCase = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(_UpperCamelCase ) ) # The ratio of the area for circle to square is pi/4. lowerCAmelCase = proportion * 4 print(F"""The estimated value of pi is {pi_estimate}""" ) print(F"""The numpy value of pi is {pi}""" ) print(F"""The total error is {abs(pi - pi_estimate )}""" ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case = 0.0 , _snake_case = 1.0 , ): '''simple docstring''' return mean( function_to_integrate(uniform(_UpperCamelCase , _UpperCamelCase ) ) for _ in range(_UpperCamelCase ) ) * (max_value - min_value) def UpperCAmelCase ( _snake_case , _snake_case = 0.0 , _snake_case = 1.0 ): '''simple docstring''' def identity_function(_snake_case ) -> float: return x lowerCAmelCase = area_under_curve_estimator( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) lowerCAmelCase = (max_value * max_value - min_value * min_value) / 2 print('''******************''' ) print(F"""Estimating area under y=x where x varies from {min_value} to {max_value}""" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {expected_value}""" ) print(F"""Total error is {abs(estimated_value - expected_value )}""" ) print('''******************''' ) def UpperCAmelCase ( _snake_case ): '''simple docstring''' def function_to_integrate(_snake_case ) -> float: return sqrt(4.0 - x * x ) lowerCAmelCase = area_under_curve_estimator( _UpperCamelCase , _UpperCamelCase , 0.0 , 2.0 ) print('''******************''' ) print('''Estimating pi using area_under_curve_estimator''' ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {pi}""" ) print(F"""Total error is {abs(estimated_value - pi )}""" ) print('''******************''' ) if __name__ == "__main__": import doctest doctest.testmod()
715
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) UpperCAmelCase_ ="""hf-internal-testing/tiny-random-bert""" UpperCAmelCase_ =os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") UpperCAmelCase_ ="""9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCAmelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertTrue(os.path.isfile(UpperCAmelCase_ ) ) # File is cached at the same place the second time. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Using a specific revision to test the full commit hash. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''9b8c223''' ) self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): lowerCAmelCase = cached_file('''tiny-random-bert''' , UpperCAmelCase_ ) with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCAmelCase_ , '''.no_exist''' , UpperCAmelCase_ , '''conf''' ) ) ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , local_files_only=UpperCAmelCase_ , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = mock.Mock() lowerCAmelCase = 5_00 lowerCAmelCase = {} lowerCAmelCase = HTTPError lowerCAmelCase = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCAmelCase_ ) as mock_head: lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) # This check we did call the fake head request mock_head.assert_called() def __snake_case ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) def __snake_case ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCAmelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ , revision='''ahaha''' ) lowerCAmelCase = get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCAmelCase = json.loads(open(UpperCAmelCase_ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_68 ) def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase = Path(UpperCAmelCase_ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCAmelCase_ , '''a.txt''' ) , str(UpperCAmelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCAmelCase_ , '''b.txt''' ) )
33
0
import os import numpy import onnx def UpperCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase = a.name lowerCAmelCase = b.name lowerCAmelCase = '''''' lowerCAmelCase = '''''' lowerCAmelCase = a == b lowerCAmelCase = name_a lowerCAmelCase = name_b return res def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(_snake_case , _snake_case ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , _snake_case , _snake_case ) _graph_replace_input_with(node_proto.attribute[1].g , _snake_case , _snake_case ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , _snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): for n in graph_proto.node: _node_replace_input_with(_snake_case , _snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = list(model.graph.initializer ) lowerCAmelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i lowerCAmelCase = inits[i].name lowerCAmelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , _snake_case , _snake_case ) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = os.path.dirname(_snake_case ) lowerCAmelCase = os.path.basename(_snake_case ) lowerCAmelCase = onnx.load(os.path.join(_snake_case , _snake_case ) ) lowerCAmelCase = list(model.graph.initializer ) lowerCAmelCase = set() lowerCAmelCase = {} lowerCAmelCase = [] lowerCAmelCase = 0 for i in range(len(_snake_case ) ): if i in dup_set: continue for j in range(i + 1 , len(_snake_case ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(_snake_case ) dup_set.add(_snake_case ) lowerCAmelCase = inits[j].data_type lowerCAmelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('''unexpected data type: ''' , _snake_case ) total_reduced_size += mem_size lowerCAmelCase = inits[i].name lowerCAmelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(_snake_case ) else: lowerCAmelCase = [name_j] ind_to_replace.append((j, i) ) print('''total reduced size: ''' , total_reduced_size / 1024 / 1024 / 1024 , '''GB''' ) lowerCAmelCase = sorted(_snake_case ) _remove_dup_initializers_from_model(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = '''optimized_''' + model_file_name lowerCAmelCase = os.path.join(_snake_case , _snake_case ) onnx.save(_snake_case , _snake_case ) return new_model
716
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = "arrow" , **UpperCAmelCase_ , ): super().__init__( split=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , keep_in_memory=UpperCAmelCase_ , streaming=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = load_from_cache_file lowerCAmelCase = file_format lowerCAmelCase = Spark( df=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , working_dir=UpperCAmelCase_ , **UpperCAmelCase_ , ) def __snake_case ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
33
0
def UpperCAmelCase ( _snake_case ): if not isinstance(_snake_case , _snake_case ): lowerCAmelCase = F"""Input value of [number={number}] must be an integer""" raise TypeError(_snake_case ) if number < 0: return False lowerCAmelCase = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
717
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase ( _snake_case = 3 ): if isinstance(_snake_case , _snake_case ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(_snake_case ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) lowerCAmelCase = QuantumRegister(_snake_case , '''qr''' ) lowerCAmelCase = ClassicalRegister(_snake_case , '''cr''' ) lowerCAmelCase = QuantumCircuit(_snake_case , _snake_case ) lowerCAmelCase = number_of_qubits for i in range(_snake_case ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(_snake_case ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , _snake_case , _snake_case ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(_snake_case , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(_snake_case , _snake_case ) # simulate with 10000 shots lowerCAmelCase = Aer.get_backend('''qasm_simulator''' ) lowerCAmelCase = execute(_snake_case , _snake_case , shots=10000 ) return job.result().get_counts(_snake_case ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
33
0
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=64 , UpperCAmelCase_=2 , UpperCAmelCase_=3 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=32 , UpperCAmelCase_=5 , UpperCAmelCase_=4 , UpperCAmelCase_=37 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=10 , UpperCAmelCase_=0.02 , UpperCAmelCase_=[1, 16, 4, 4] , UpperCAmelCase_=None , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = scope lowerCAmelCase = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size lowerCAmelCase = (self.image_size // 32) ** 2 lowerCAmelCase = num_patches + 1 def __snake_case ( self ): lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = self.get_config() return config, pixel_values, labels def __snake_case ( self ): lowerCAmelCase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [4, 8, 16, 32], '''num_groups''': 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase_ , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=UpperCAmelCase_ , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = ViTHybridModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = self.type_sequence_label_size lowerCAmelCase = ViTHybridForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() lowerCAmelCase = config_and_inputs lowerCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): '''simple docstring''' __a : Any =(ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __a : Dict =( {"""feature-extraction""": ViTHybridModel, """image-classification""": ViTHybridForImageClassification} if is_torch_available() else {} ) __a : str =False __a : Optional[int] =False __a : Union[str, Any] =False def __snake_case ( self ): lowerCAmelCase = ViTHybridModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ , hidden_size=37 ) def __snake_case ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def __snake_case ( self ): pass def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(UpperCAmelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase_ , nn.Linear ) ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(UpperCAmelCase_ ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = _config_zero_init(UpperCAmelCase_ ) for model_class in self.all_model_classes: lowerCAmelCase = model_class(config=UpperCAmelCase_ ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": lowerCAmelCase = [F"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @slow def __snake_case ( self ): for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = ViTHybridModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCAmelCase ( ): lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __snake_case ( self ): return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self ): lowerCAmelCase = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( UpperCAmelCase_ ) lowerCAmelCase = self.default_image_processor lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**UpperCAmelCase_ ) # verify the logits lowerCAmelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow @require_accelerate def __snake_case ( self ): lowerCAmelCase = ViTHybridImageProcessor.from_pretrained('''google/vit-hybrid-base-bit-384''' ) lowerCAmelCase = ViTHybridForImageClassification.from_pretrained('''google/vit-hybrid-base-bit-384''' , device_map='''auto''' ) lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ) lowerCAmelCase = model(**UpperCAmelCase_ ) lowerCAmelCase = outputs.logits # model predicts one of the 1000 ImageNet classes lowerCAmelCase = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , '''tabby, tabby cat''' )
718
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Any =1 @register_to_config def __init__( self , UpperCAmelCase_=20_00 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=1E-3 ): lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = torch.linspace(1 , self.config.sampling_eps , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCAmelCase = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCAmelCase = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCAmelCase = std.flatten() while len(std.shape ) < len(score.shape ): lowerCAmelCase = std.unsqueeze(-1 ) lowerCAmelCase = -score / std # compute lowerCAmelCase = -1.0 / len(self.timesteps ) lowerCAmelCase = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCAmelCase = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCAmelCase = beta_t.unsqueeze(-1 ) lowerCAmelCase = -0.5 * beta_t * x lowerCAmelCase = torch.sqrt(UpperCAmelCase_ ) lowerCAmelCase = drift - diffusion**2 * score lowerCAmelCase = x + drift * dt # add noise lowerCAmelCase = randn_tensor(x.shape , layout=x.layout , generator=UpperCAmelCase_ , device=x.device , dtype=x.dtype ) lowerCAmelCase = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ): return self.config.num_train_timesteps
33
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class __UpperCamelCase ( lowercase_ ): '''simple docstring''' __a : Optional[int] ='''deta''' __a : str ={ '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , UpperCAmelCase_=None , UpperCAmelCase_=9_00 , UpperCAmelCase_=20_48 , UpperCAmelCase_=6 , UpperCAmelCase_=20_48 , UpperCAmelCase_=8 , UpperCAmelCase_=6 , UpperCAmelCase_=10_24 , UpperCAmelCase_=8 , UpperCAmelCase_=0.0 , UpperCAmelCase_=True , UpperCAmelCase_="relu" , UpperCAmelCase_=2_56 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.02 , UpperCAmelCase_=1.0 , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_="sine" , UpperCAmelCase_=5 , UpperCAmelCase_=4 , UpperCAmelCase_=4 , UpperCAmelCase_=True , UpperCAmelCase_=3_00 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=1 , UpperCAmelCase_=5 , UpperCAmelCase_=2 , UpperCAmelCase_=1 , UpperCAmelCase_=1 , UpperCAmelCase_=5 , UpperCAmelCase_=2 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.25 , **UpperCAmelCase_ , ): if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) lowerCAmelCase = CONFIG_MAPPING['''resnet'''](out_features=['''stage2''', '''stage3''', '''stage4'''] ) else: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = backbone_config.pop('''model_type''' ) lowerCAmelCase = CONFIG_MAPPING[backbone_model_type] lowerCAmelCase = config_class.from_dict(UpperCAmelCase_ ) lowerCAmelCase = backbone_config lowerCAmelCase = num_queries lowerCAmelCase = max_position_embeddings lowerCAmelCase = d_model lowerCAmelCase = encoder_ffn_dim lowerCAmelCase = encoder_layers lowerCAmelCase = encoder_attention_heads lowerCAmelCase = decoder_ffn_dim lowerCAmelCase = decoder_layers lowerCAmelCase = decoder_attention_heads lowerCAmelCase = dropout lowerCAmelCase = attention_dropout lowerCAmelCase = activation_dropout lowerCAmelCase = activation_function lowerCAmelCase = init_std lowerCAmelCase = init_xavier_std lowerCAmelCase = encoder_layerdrop lowerCAmelCase = auxiliary_loss lowerCAmelCase = position_embedding_type # deformable attributes lowerCAmelCase = num_feature_levels lowerCAmelCase = encoder_n_points lowerCAmelCase = decoder_n_points lowerCAmelCase = two_stage lowerCAmelCase = two_stage_num_proposals lowerCAmelCase = with_box_refine lowerCAmelCase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''' ) # Hungarian matcher lowerCAmelCase = class_cost lowerCAmelCase = bbox_cost lowerCAmelCase = giou_cost # Loss coefficients lowerCAmelCase = mask_loss_coefficient lowerCAmelCase = dice_loss_coefficient lowerCAmelCase = bbox_loss_coefficient lowerCAmelCase = giou_loss_coefficient lowerCAmelCase = eos_coefficient lowerCAmelCase = focal_alpha super().__init__(is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ ) @property def __snake_case ( self ): return self.encoder_attention_heads @property def __snake_case ( self ): return self.d_model def __snake_case ( self ): lowerCAmelCase = copy.deepcopy(self.__dict__ ) lowerCAmelCase = self.backbone_config.to_dict() lowerCAmelCase = self.__class__.model_type return output
719
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
33
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """xlm-mlm-en-2048""": """https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json""", """xlm-mlm-ende-1024""": """https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json""", """xlm-mlm-enfr-1024""": """https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json""", """xlm-mlm-enro-1024""": """https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json""", """xlm-mlm-tlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json""", """xlm-mlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json""", """xlm-clm-enfr-1024""": """https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json""", """xlm-clm-ende-1024""": """https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json""", """xlm-mlm-17-1280""": """https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json""", """xlm-mlm-100-1280""": """https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : int ="""xlm""" __a : Dict ={ """hidden_size""": """emb_dim""", """num_attention_heads""": """n_heads""", """num_hidden_layers""": """n_layers""", """n_words""": """vocab_size""", # For backward compatibility } def __init__( self , UpperCAmelCase_=3_01_45 , UpperCAmelCase_=20_48 , UpperCAmelCase_=12 , UpperCAmelCase_=16 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=False , UpperCAmelCase_=False , UpperCAmelCase_=1 , UpperCAmelCase_=True , UpperCAmelCase_=5_12 , UpperCAmelCase_=20_48**-0.5 , UpperCAmelCase_=1E-1_2 , UpperCAmelCase_=0.02 , UpperCAmelCase_=0 , UpperCAmelCase_=1 , UpperCAmelCase_=2 , UpperCAmelCase_=3 , UpperCAmelCase_=5 , UpperCAmelCase_=True , UpperCAmelCase_="first" , UpperCAmelCase_=True , UpperCAmelCase_=None , UpperCAmelCase_=True , UpperCAmelCase_=0.1 , UpperCAmelCase_=5 , UpperCAmelCase_=5 , UpperCAmelCase_=0 , UpperCAmelCase_=0 , UpperCAmelCase_=2 , UpperCAmelCase_=0 , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = emb_dim lowerCAmelCase = n_layers lowerCAmelCase = n_heads lowerCAmelCase = dropout lowerCAmelCase = attention_dropout lowerCAmelCase = gelu_activation lowerCAmelCase = sinusoidal_embeddings lowerCAmelCase = causal lowerCAmelCase = asm lowerCAmelCase = n_langs lowerCAmelCase = use_lang_emb lowerCAmelCase = layer_norm_eps lowerCAmelCase = bos_index lowerCAmelCase = eos_index lowerCAmelCase = pad_index lowerCAmelCase = unk_index lowerCAmelCase = mask_index lowerCAmelCase = is_encoder lowerCAmelCase = max_position_embeddings lowerCAmelCase = embed_init_std lowerCAmelCase = init_std lowerCAmelCase = summary_type lowerCAmelCase = summary_use_proj lowerCAmelCase = summary_activation lowerCAmelCase = summary_proj_to_labels lowerCAmelCase = summary_first_dropout lowerCAmelCase = start_n_top lowerCAmelCase = end_n_top lowerCAmelCase = mask_token_id lowerCAmelCase = lang_id if "n_words" in kwargs: lowerCAmelCase = kwargs['''n_words'''] super().__init__(pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' @property def __snake_case ( self ): if self.task == "multiple-choice": lowerCAmelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
720
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : str =MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = VideoClassificationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ , top_k=2 ) lowerCAmelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): for example in examples: lowerCAmelCase = video_classifier(UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, ] , ) @require_torch def __snake_case ( self ): lowerCAmelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' lowerCAmelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) lowerCAmelCase = pipeline( '''video-classification''' , model=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = video_classifier(UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __snake_case ( self ): pass
33
0
from __future__ import annotations from collections.abc import Iterator class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = value lowerCAmelCase = None lowerCAmelCase = None class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = tree def __snake_case ( self , UpperCAmelCase_ ): if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ): yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
721
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any ="""hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __snake_case ( self , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCAmelCase_ ) ) lowerCAmelCase = np.random.RandomState(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # warmup pass to apply optimizations lowerCAmelCase = pipe(**self.get_dummy_inputs() ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @property def __snake_case ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __snake_case ( self ): lowerCAmelCase = ort.SessionOptions() lowerCAmelCase = False return options def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCAmelCase_ , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
33
0
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def UpperCAmelCase ( _snake_case ): return EnvironmentCommand() def UpperCAmelCase ( _snake_case ): return EnvironmentCommand(args.accelerate_config_file ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' @staticmethod def __snake_case ( UpperCAmelCase_ ): lowerCAmelCase = parser.add_parser('''env''' ) download_parser.set_defaults(func=UpperCAmelCase_ ) download_parser.add_argument( '''--accelerate-config_file''' , default=UpperCAmelCase_ , help='''The accelerate config file to use for the default values in the launching script.''' , ) download_parser.set_defaults(func=UpperCAmelCase_ ) def __init__( self , UpperCAmelCase_ , *UpperCAmelCase_ ): lowerCAmelCase = accelerate_config_file def __snake_case ( self ): lowerCAmelCase = '''not installed''' if is_safetensors_available(): import safetensors lowerCAmelCase = safetensors.__version__ elif importlib.util.find_spec('''safetensors''' ) is not None: import safetensors lowerCAmelCase = F"""{safetensors.__version__} but is ignored because of PyTorch version too old.""" lowerCAmelCase = '''not installed''' lowerCAmelCase = lowerCAmelCase = '''not found''' if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file lowerCAmelCase = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(UpperCAmelCase_ ): lowerCAmelCase = load_config_from_file(self._accelerate_config_file ).to_dict() lowerCAmelCase = ( '''\n'''.join([F"""\t- {prop}: {val}""" for prop, val in accelerate_config.items()] ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else F"""\t{accelerate_config}""" ) lowerCAmelCase = '''not installed''' lowerCAmelCase = '''NA''' if is_torch_available(): import torch lowerCAmelCase = torch.__version__ lowerCAmelCase = torch.cuda.is_available() lowerCAmelCase = '''not installed''' lowerCAmelCase = '''NA''' if is_tf_available(): import tensorflow as tf lowerCAmelCase = tf.__version__ try: # deprecated in v2.1 lowerCAmelCase = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool lowerCAmelCase = bool(tf.config.list_physical_devices('''GPU''' ) ) lowerCAmelCase = '''not installed''' lowerCAmelCase = '''not installed''' lowerCAmelCase = '''not installed''' lowerCAmelCase = '''NA''' if is_flax_available(): import flax import jax import jaxlib lowerCAmelCase = flax.__version__ lowerCAmelCase = jax.__version__ lowerCAmelCase = jaxlib.__version__ lowerCAmelCase = jax.lib.xla_bridge.get_backend().platform lowerCAmelCase = { '''`transformers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Huggingface_hub version''': huggingface_hub.__version__, '''Safetensors version''': F"""{safetensors_version}""", '''Accelerate version''': F"""{accelerate_version}""", '''Accelerate config''': F"""{accelerate_config_str}""", '''PyTorch version (GPU?)''': F"""{pt_version} ({pt_cuda_available})""", '''Tensorflow version (GPU?)''': F"""{tf_version} ({tf_cuda_available})""", '''Flax version (CPU?/GPU?/TPU?)''': F"""{flax_version} ({jax_backend})""", '''Jax version''': F"""{jax_version}""", '''JaxLib version''': F"""{jaxlib_version}""", '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(UpperCAmelCase_ ) ) return info @staticmethod def __snake_case ( UpperCAmelCase_ ): return "\n".join([F"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
700
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCAmelCase ( _snake_case ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('''/''' ) lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase = torch.load(os.path.join(_snake_case , '''pytorch_model.bin''' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase , lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_snake_case ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F"""bertarized_{os.path.basename(_snake_case )}""" ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(_snake_case , os.path.join(_snake_case , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) UpperCAmelCase_ =parser.parse_args() main(args)
33
0
def UpperCAmelCase ( _snake_case ): if num <= 0: raise ValueError('''Input must be a positive integer''' ) lowerCAmelCase = [True] * (num + 1) lowerCAmelCase = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , _snake_case ): lowerCAmelCase = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =int(input("""Enter a positive integer: """).strip()) print(prime_sieve_eratosthenes(user_num))
701
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } UpperCAmelCase_ ={ """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } UpperCAmelCase_ ={ """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = set() lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase = char lowerCAmelCase = set(_snake_case ) return pairs class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =VOCAB_FILES_NAMES __a : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP __a : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , **UpperCAmelCase_ , ): super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = merges_file lowerCAmelCase = {} lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 lowerCAmelCase = 3 self.add_from_file(UpperCAmelCase_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase_ , encoding='''utf-8''' ) as merges_handle: lowerCAmelCase = merges_handle.read().split('''\n''' )[:-1] lowerCAmelCase = [tuple(merge.split()[:-1] ) for merge in merges] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __snake_case ( self ): return len(self.encoder ) def __snake_case ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self , UpperCAmelCase_ ): if token in self.cache: return self.cache[token] lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowerCAmelCase = get_pairs(UpperCAmelCase_ ) if not pairs: return token while True: lowerCAmelCase = min(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : self.bpe_ranks.get(UpperCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase = bigram lowerCAmelCase = [] lowerCAmelCase = 0 while i < len(UpperCAmelCase_ ): try: lowerCAmelCase = word.index(UpperCAmelCase_ , UpperCAmelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase = j if word[i] == first and i < len(UpperCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = new_word if len(UpperCAmelCase_ ) == 1: break else: lowerCAmelCase = get_pairs(UpperCAmelCase_ ) lowerCAmelCase = '''@@ '''.join(UpperCAmelCase_ ) lowerCAmelCase = word[:-4] lowerCAmelCase = word return word def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [] lowerCAmelCase = re.findall(r'''\S+\n?''' , UpperCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __snake_case ( self , UpperCAmelCase_ ): return self.encoder.get(UpperCAmelCase_ , self.encoder.get(self.unk_token ) ) def __snake_case ( self , UpperCAmelCase_ ): return self.decoder.get(UpperCAmelCase_ , self.unk_token ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = ''' '''.join(UpperCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.merges_file , UpperCAmelCase_ ) return out_vocab_file, out_merge_file def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(UpperCAmelCase_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCAmelCase = f.readlines() for lineTmp in lines: lowerCAmelCase = lineTmp.strip() lowerCAmelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowerCAmelCase = line[:idx] lowerCAmelCase = len(self.encoder )
33
0
def UpperCAmelCase ( _snake_case ): lowerCAmelCase = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def UpperCAmelCase ( _snake_case = 100 ): lowerCAmelCase = 1 lowerCAmelCase = 2 for i in range(2 , max_n + 1 ): lowerCAmelCase = pre_numerator lowerCAmelCase = 2 * i // 3 if i % 3 == 0 else 1 lowerCAmelCase = cur_numerator lowerCAmelCase = e_cont * pre_numerator + temp return sum_digits(_snake_case ) if __name__ == "__main__": print(F'''{solution() = }''')
702
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ =TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = self lowerCAmelCase = 0 class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # create a new set with x as its member lowerCAmelCase = DisjointSetTreeNode(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): # find the set x belongs to (with path-compression) lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # helper function for union operation if nodea.rank > nodea.rank: lowerCAmelCase = nodea else: lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # merge 2 disjoint sets self.link(self.find_set(UpperCAmelCase_ ) , self.find_set(UpperCAmelCase_ ) ) class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # add a node ONLY if its not present in the graph if node not in self.connections: lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # add an edge with the given weight self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) lowerCAmelCase = weight lowerCAmelCase = weight def __snake_case ( self ): lowerCAmelCase = [] lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCAmelCase_ : x[2] ) # creating the disjoint set lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(UpperCAmelCase_ ) # MST generation lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = edges[index] index += 1 lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) disjoint_set.union(UpperCAmelCase_ , UpperCAmelCase_ ) return graph
33
0
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =IFImgaImgSuperResolutionPipeline __a : str =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} __a : Tuple =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) __a : Any =PipelineTesterMixin.required_optional_params - {"""latents"""} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
703
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations(_snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations_with_dp_array( _snake_case , _snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _snake_case ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =3 UpperCAmelCase_ =5 UpperCAmelCase_ =[1, 2, 5] print(combination_sum_iv(n, array, target))
33
0
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def UpperCAmelCase ( _snake_case = "laptop" ): lowerCAmelCase = F"""https://www.amazon.in/laptop/s?k={product}""" lowerCAmelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } lowerCAmelCase = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles lowerCAmelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: lowerCAmelCase = item.ha.text lowerCAmelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] lowerCAmelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: lowerCAmelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: lowerCAmelCase = '''Not available''' try: lowerCAmelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: lowerCAmelCase = '''''' try: lowerCAmelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 100 ) except ValueError: lowerCAmelCase = float('''nan''' ) except AttributeError: pass lowerCAmelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] lowerCAmelCase = ''' ''' lowerCAmelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": UpperCAmelCase_ ="""headphones""" get_amazon_product_data(product).to_csv(F'''Amazon Product Data for {product}.csv''')
704
import torch from diffusers import StableDiffusionPipeline UpperCAmelCase_ ="""path-to-your-trained-model""" UpperCAmelCase_ =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") UpperCAmelCase_ ="""A photo of sks dog in a bucket""" UpperCAmelCase_ =pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
33
0
'''simple docstring''' def UpperCAmelCase ( _snake_case ): if not isinstance(_snake_case , _snake_case ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer UpperCAmelCase_ ="""bart""" UpperCAmelCase_ =True @st.cache(allow_output_mutation=_snake_case ) def UpperCAmelCase ( ): if LOAD_DENSE_INDEX: lowerCAmelCase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) lowerCAmelCase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) lowerCAmelCase = qar_model.eval() else: lowerCAmelCase , lowerCAmelCase = (None, None) if MODEL_TYPE == "bart": lowerCAmelCase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) lowerCAmelCase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) lowerCAmelCase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) lowerCAmelCase = sas_model.eval() else: lowerCAmelCase , lowerCAmelCase = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=_snake_case ) def UpperCAmelCase ( ): if LOAD_DENSE_INDEX: lowerCAmelCase = faiss.StandardGpuResources() lowerCAmelCase = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] lowerCAmelCase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 128) , ) lowerCAmelCase = faiss.IndexFlatIP(128 ) lowerCAmelCase = faiss.index_cpu_to_gpu(_snake_case , 1 , _snake_case ) wikiaab_gpu_index_flat.add(_snake_case ) # TODO fix for larger GPU else: lowerCAmelCase , lowerCAmelCase = (None, None) lowerCAmelCase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=_snake_case ) def UpperCAmelCase ( ): lowerCAmelCase = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) lowerCAmelCase = elia['''train_eli5'''] lowerCAmelCase = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 128) ) lowerCAmelCase = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(_snake_case ) return (elia_train, eli5_train_q_index) UpperCAmelCase_,UpperCAmelCase_,UpperCAmelCase_ =load_indexes() UpperCAmelCase_,UpperCAmelCase_,UpperCAmelCase_,UpperCAmelCase_ =load_models() UpperCAmelCase_,UpperCAmelCase_ =load_train_data() def UpperCAmelCase ( _snake_case , _snake_case=10 ): lowerCAmelCase = embed_questions_for_retrieval([question] , _snake_case , _snake_case ) lowerCAmelCase , lowerCAmelCase = eli5_train_q_index.search(_snake_case , _snake_case ) lowerCAmelCase = [elia_train[int(_snake_case )] for i in I[0]] return nn_examples def UpperCAmelCase ( _snake_case , _snake_case="wiki40b" , _snake_case="dense" , _snake_case=10 ): if source == "none": lowerCAmelCase , lowerCAmelCase = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": lowerCAmelCase , lowerCAmelCase = query_qa_dense_index( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) else: lowerCAmelCase , lowerCAmelCase = query_es_index( _snake_case , _snake_case , index_name='''english_wiki40b_snippets_100w''' , n_results=_snake_case , ) lowerCAmelCase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] lowerCAmelCase = '''question: {} context: {}'''.format(_snake_case , _snake_case ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda _snake_case : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda _snake_case : None), } ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case=64 , _snake_case=256 , _snake_case=False , _snake_case=2 , _snake_case=0.95 , _snake_case=0.8 ): with torch.no_grad(): lowerCAmelCase = qa_sas_generate( _snake_case , _snake_case , _snake_case , num_answers=1 , num_beams=_snake_case , min_len=_snake_case , max_len=_snake_case , do_sample=_snake_case , temp=_snake_case , top_p=_snake_case , top_k=_snake_case , max_input_length=1024 , device='''cuda:0''' , )[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar UpperCAmelCase_ ="""<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" UpperCAmelCase_ =""" <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia UpperCAmelCase_ =""" This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) UpperCAmelCase_ =[ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] UpperCAmelCase_ =st.sidebar.checkbox("""Demo options""") if demo_options: UpperCAmelCase_ =st.sidebar.selectbox( """""", action_list, index=3, ) UpperCAmelCase_ =action_list.index(action_st) UpperCAmelCase_ =st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) UpperCAmelCase_ =show_type == """Show full text of passages""" else: UpperCAmelCase_ =3 UpperCAmelCase_ =True UpperCAmelCase_ =st.sidebar.checkbox("""Retrieval options""") if retrieval_options: UpperCAmelCase_ =""" ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) UpperCAmelCase_ =st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) UpperCAmelCase_ =st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: UpperCAmelCase_ ="""wiki40b""" UpperCAmelCase_ ="""dense""" UpperCAmelCase_ ="""beam""" UpperCAmelCase_ =2 UpperCAmelCase_ =64 UpperCAmelCase_ =256 UpperCAmelCase_ =None UpperCAmelCase_ =None UpperCAmelCase_ =st.sidebar.checkbox("""Generation options""") if generate_options: UpperCAmelCase_ =""" ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) UpperCAmelCase_ =st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) UpperCAmelCase_ =st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) UpperCAmelCase_ =st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": UpperCAmelCase_ =st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: UpperCAmelCase_ =st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) UpperCAmelCase_ =st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) UpperCAmelCase_ =None # start main text UpperCAmelCase_ =[ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] UpperCAmelCase_ =st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": UpperCAmelCase_ =st.text_input("""Enter your question here:""", """""") else: UpperCAmelCase_ =question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": UpperCAmelCase_,UpperCAmelCase_ =make_support(question, source=wiki_source, method="""dense""", n_results=10) UpperCAmelCase_,UpperCAmelCase_ =make_support(question, source=wiki_source, method="""sparse""", n_results=10) UpperCAmelCase_ =[] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] UpperCAmelCase_ =support_list[:10] UpperCAmelCase_ ="""<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: UpperCAmelCase_,UpperCAmelCase_ =make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: UpperCAmelCase_,UpperCAmelCase_ =answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): UpperCAmelCase_ ="""https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) UpperCAmelCase_ =res[1].strip() if sec_titles == "": UpperCAmelCase_ ="""[{}]({})""".format(res[0], wiki_url) else: UpperCAmelCase_ =sec_titles.split(""" & """) UpperCAmelCase_ =""" & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: UpperCAmelCase_ =find_nearest_training(question) UpperCAmelCase_ =nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) UpperCAmelCase_ =[ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) UpperCAmelCase_ =""" --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
706
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = 8 # DPR tok lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __snake_case ( self ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __snake_case ( self ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: lowerCAmelCase = os.path.join(self.tmpdirname , '''dataset''' ) lowerCAmelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCAmelCase_ ) , ) return retriever def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) lowerCAmelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) lowerCAmelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(UpperCAmelCase_ , open(UpperCAmelCase_ , '''wb''' ) ) lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = self.get_dummy_dataset() retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_legacy_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): import torch lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) lowerCAmelCase = retriever( UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ , return_tensors='''pt''' , ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): lowerCAmelCase = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) retriever.set_ctx_encoder_tokenizer(UpperCAmelCase_ ) lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) self.assertEqual( len(UpperCAmelCase_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCAmelCase_ ) # check for doc token related keys in dictionary.
33
0
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Optional[int] =OpenAIGPTTokenizer __a : Tuple =OpenAIGPTTokenizerFast __a : Any =True __a : Union[str, Any] =False def __snake_case ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self , UpperCAmelCase_ ): return "lower newer", "lower newer" def __snake_case ( self ): lowerCAmelCase = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase = '''lower''' lowerCAmelCase = ['''low''', '''er</w>'''] lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = tokens + ['''<unk>'''] lowerCAmelCase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) # Simple input lowerCAmelCase = '''This is a simple input''' lowerCAmelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] lowerCAmelCase = ('''This is a simple input''', '''This is a pair''') lowerCAmelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase_ , tokenizer_r.encode , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase_ , tokenizer_r.encode_plus , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase_ , tokenizer_r.batch_encode_plus , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase_ , tokenizer_r.encode , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase_ , tokenizer_r.encode_plus , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase_ , tokenizer_r.batch_encode_plus , UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding='''max_length''' , ) def __snake_case ( self ): pass @require_ftfy @require_spacy @require_tokenizers class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' pass
707
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """google/switch-base-8""": """https://huggingface.co/google/switch-base-8/blob/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ="""switch_transformers""" __a : Union[str, Any] =["""past_key_values"""] __a : Dict ={"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , UpperCAmelCase_=3_21_28 , UpperCAmelCase_=7_68 , UpperCAmelCase_=64 , UpperCAmelCase_=20_48 , UpperCAmelCase_=64 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=8 , UpperCAmelCase_=False , UpperCAmelCase_=0.01 , UpperCAmelCase_="float32" , UpperCAmelCase_=False , UpperCAmelCase_=32 , UpperCAmelCase_=1_28 , UpperCAmelCase_=0.1 , UpperCAmelCase_=1E-6 , UpperCAmelCase_=0.001 , UpperCAmelCase_=0.001 , UpperCAmelCase_=1.0 , UpperCAmelCase_="relu" , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=0 , UpperCAmelCase_=1 , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = d_kv lowerCAmelCase = d_ff lowerCAmelCase = num_sparse_encoder_layers lowerCAmelCase = num_layers lowerCAmelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: lowerCAmelCase = self.num_layers // self.num_sparse_encoder_layers else: lowerCAmelCase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: lowerCAmelCase = self.num_decoder_layers // self.num_sparse_decoder_layers else: lowerCAmelCase = self.num_decoder_layers # HACK: this will create 0 sparse layers lowerCAmelCase = num_heads lowerCAmelCase = num_experts lowerCAmelCase = expert_capacity lowerCAmelCase = router_bias lowerCAmelCase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) lowerCAmelCase = router_dtype lowerCAmelCase = router_ignore_padding_tokens lowerCAmelCase = relative_attention_num_buckets lowerCAmelCase = relative_attention_max_distance lowerCAmelCase = dropout_rate lowerCAmelCase = layer_norm_epsilon lowerCAmelCase = initializer_factor lowerCAmelCase = feed_forward_proj lowerCAmelCase = use_cache lowerCAmelCase = add_router_probs lowerCAmelCase = router_z_loss_coef lowerCAmelCase = router_aux_loss_coef lowerCAmelCase = self.feed_forward_proj.split('''-''' ) lowerCAmelCase = act_info[-1] lowerCAmelCase = act_info[0] == '''gated''' if len(UpperCAmelCase_ ) > 1 and act_info[0] != "gated" or len(UpperCAmelCase_ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCAmelCase = '''gelu_new''' super().__init__( pad_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ , )
33
0
import collections import importlib.util import os import re from pathlib import Path UpperCAmelCase_ ="""src/transformers""" # Matches is_xxx_available() UpperCAmelCase_ =re.compile(R"""is\_([a-z_]*)_available()""") # Catches a one-line _import_struct = {xxx} UpperCAmelCase_ =re.compile(R"""^_import_structure\s+=\s+\{([^\}]+)\}""") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] UpperCAmelCase_ =re.compile(R"""\s+\"\S*\":\s+\[([^\]]*)\]""") # Catches a line if not is_foo_available UpperCAmelCase_ =re.compile(R"""^\s*if\s+not\s+is\_[a-z_]*\_available\(\)""") # Catches a line _import_struct["bla"].append("foo") UpperCAmelCase_ =re.compile(R"""^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)""") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] UpperCAmelCase_ =re.compile(R"""^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]""") # Catches a line with an object between quotes and a comma: "MyModel", UpperCAmelCase_ =re.compile("""^\s+\"([^\"]+)\",""") # Catches a line with objects between brackets only: ["foo", "bar"], UpperCAmelCase_ =re.compile("""^\s+\[([^\]]+)\]""") # Catches a line with from foo import bar, bla, boo UpperCAmelCase_ =re.compile(R"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") # Catches a line with try: UpperCAmelCase_ =re.compile(R"""^\s*try:""") # Catches a line with else: UpperCAmelCase_ =re.compile(R"""^\s*else:""") def UpperCAmelCase ( _snake_case ): if _re_test_backend.search(_snake_case ) is None: return None lowerCAmelCase = [b[0] for b in _re_backend.findall(_snake_case )] backends.sort() return "_and_".join(_snake_case ) def UpperCAmelCase ( _snake_case ): with open(_snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase = f.readlines() lowerCAmelCase = 0 while line_index < len(_snake_case ) and not lines[line_index].startswith('''_import_structure = {''' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(_snake_case ): return None # First grab the objects without a specific backend in _import_structure lowerCAmelCase = [] while not lines[line_index].startswith('''if TYPE_CHECKING''' ) and find_backend(lines[line_index] ) is None: lowerCAmelCase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(_snake_case ): lowerCAmelCase = _re_one_line_import_struct.search(_snake_case ).groups()[0] lowerCAmelCase = re.findall('''\[([^\]]+)\]''' , _snake_case ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(''', ''' )] ) line_index += 1 continue lowerCAmelCase = _re_import_struct_key_value.search(_snake_case ) if single_line_import_search is not None: lowerCAmelCase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(''', ''' ) if len(_snake_case ) > 0] objects.extend(_snake_case ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) line_index += 1 lowerCAmelCase = {'''none''': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('''if TYPE_CHECKING''' ): # If the line is an if not is_backend_available, we grab all objects associated. lowerCAmelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowerCAmelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowerCAmelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 4 ): lowerCAmelCase = lines[line_index] if _re_import_struct_add_one.search(_snake_case ) is not None: objects.append(_re_import_struct_add_one.search(_snake_case ).groups()[0] ) elif _re_import_struct_add_many.search(_snake_case ) is not None: lowerCAmelCase = _re_import_struct_add_many.search(_snake_case ).groups()[0].split(''', ''' ) lowerCAmelCase = [obj[1:-1] for obj in imports if len(_snake_case ) > 0] objects.extend(_snake_case ) elif _re_between_brackets.search(_snake_case ) is not None: lowerCAmelCase = _re_between_brackets.search(_snake_case ).groups()[0].split(''', ''' ) lowerCAmelCase = [obj[1:-1] for obj in imports if len(_snake_case ) > 0] objects.extend(_snake_case ) elif _re_quote_object.search(_snake_case ) is not None: objects.append(_re_quote_object.search(_snake_case ).groups()[0] ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) elif line.startswith(''' ''' * 12 + '''"''' ): objects.append(line[13:-3] ) line_index += 1 lowerCAmelCase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend lowerCAmelCase = [] while ( line_index < len(_snake_case ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('''else''' ) ): lowerCAmelCase = lines[line_index] lowerCAmelCase = _re_import.search(_snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 lowerCAmelCase = {'''none''': objects} # Let's continue with backend-specific objects while line_index < len(_snake_case ): # If the line is an if is_backend_available, we grab all objects associated. lowerCAmelCase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowerCAmelCase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowerCAmelCase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 8 ): lowerCAmelCase = lines[line_index] lowerCAmelCase = _re_import.search(_snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 12 ): objects.append(line[12:-2] ) line_index += 1 lowerCAmelCase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def UpperCAmelCase ( _snake_case , _snake_case ): def find_duplicates(_snake_case ): return [k for k, v in collections.Counter(_snake_case ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] lowerCAmelCase = [] for key in import_dict_objects.keys(): lowerCAmelCase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F"""Duplicate _import_structure definitions for: {duplicate_imports}""" ) lowerCAmelCase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F"""Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}""" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): lowerCAmelCase = '''base imports''' if key == '''none''' else F"""{key} backend""" errors.append(F"""Differences for {name}:""" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F""" {a} in TYPE_HINT but not in _import_structure.""" ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F""" {a} in _import_structure but not in TYPE_HINT.""" ) return errors def UpperCAmelCase ( ): lowerCAmelCase = [] for root, _, files in os.walk(_snake_case ): if "__init__.py" in files: lowerCAmelCase = os.path.join(_snake_case , '''__init__.py''' ) lowerCAmelCase = parse_init(_snake_case ) if objects is not None: lowerCAmelCase = analyze_results(*_snake_case ) if len(_snake_case ) > 0: lowerCAmelCase = F"""Problem in {fname}, both halves do not define the same objects.\n{errors[0]}""" failures.append('''\n'''.join(_snake_case ) ) if len(_snake_case ) > 0: raise ValueError('''\n\n'''.join(_snake_case ) ) def UpperCAmelCase ( ): lowerCAmelCase = [] for path, directories, files in os.walk(_snake_case ): for folder in directories: # Ignore private modules if folder.startswith('''_''' ): directories.remove(_snake_case ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(_snake_case ) / folder).glob('''*.py''' ) ) ) == 0: continue lowerCAmelCase = str((Path(_snake_case ) / folder).relative_to(_snake_case ) ) lowerCAmelCase = short_path.replace(os.path.sep , '''.''' ) submodules.append(_snake_case ) for fname in files: if fname == "__init__.py": continue lowerCAmelCase = str((Path(_snake_case ) / fname).relative_to(_snake_case ) ) lowerCAmelCase = short_path.replace('''.py''' , '''''' ).replace(os.path.sep , '''.''' ) if len(submodule.split('''.''' ) ) == 1: submodules.append(_snake_case ) return submodules UpperCAmelCase_ =[ """convert_pytorch_checkpoint_to_tf2""", """modeling_flax_pytorch_utils""", ] def UpperCAmelCase ( ): # This is to make sure the transformers module imported is the one in the repo. lowerCAmelCase = importlib.util.spec_from_file_location( '''transformers''' , os.path.join(_snake_case , '''__init__.py''' ) , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) lowerCAmelCase = spec.loader.load_module() lowerCAmelCase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(_snake_case ) > 0: lowerCAmelCase = '''\n'''.join(F"""- {module}""" for module in module_not_registered ) raise ValueError( '''The following submodules are not properly registered in the main init of Transformers:\n''' F"""{list_of_modules}\n""" '''Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.''' ) if __name__ == "__main__": check_all_inits() check_submodules()
708
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def UpperCAmelCase ( _snake_case ): lowerCAmelCase , lowerCAmelCase = analyze_text(_snake_case ) lowerCAmelCase = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase = sum(single_char_strings.values() ) # one length string lowerCAmelCase = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase = single_char_strings[ch] lowerCAmelCase = my_str / all_sum my_fir_sum += prob * math.loga(_snake_case ) # entropy formula. # print entropy print(F"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase = sum(two_char_strings.values() ) lowerCAmelCase = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase = cha + cha if sequence in two_char_strings: lowerCAmelCase = two_char_strings[sequence] lowerCAmelCase = int(_snake_case ) / all_sum my_sec_sum += prob * math.loga(_snake_case ) # print second entropy print(F"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(F"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = Counter() # type: ignore lowerCAmelCase = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(_snake_case ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def UpperCAmelCase ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
33
0
from __future__ import annotations from typing import Any class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' pass class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = None def __iter__( self ): lowerCAmelCase = self lowerCAmelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(UpperCAmelCase_ ) yield node.data lowerCAmelCase = node.next_node @property def __snake_case ( self ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": UpperCAmelCase_ =Node(1) UpperCAmelCase_ =Node(2) UpperCAmelCase_ =Node(3) UpperCAmelCase_ =Node(4) print(root_node.has_loop) # False UpperCAmelCase_ =root_node.next_node print(root_node.has_loop) # True UpperCAmelCase_ =Node(5) UpperCAmelCase_ =Node(6) UpperCAmelCase_ =Node(5) UpperCAmelCase_ =Node(6) print(root_node.has_loop) # False UpperCAmelCase_ =Node(1) print(root_node.has_loop) # False
709
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Tuple =IFInpaintingSuperResolutionPipeline __a : Dict =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __a : int =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) __a : Union[str, Any] =PipelineTesterMixin.required_optional_params - {"""latents"""} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
33
0
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=3 , UpperCAmelCase_=32 , UpperCAmelCase_=3 , UpperCAmelCase_=10 , UpperCAmelCase_=[8, 16, 32, 64] , UpperCAmelCase_=[1, 1, 2, 1] , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_="relu" , UpperCAmelCase_=3 , UpperCAmelCase_=None , UpperCAmelCase_=["stage2", "stage3", "stage4"] , UpperCAmelCase_=[2, 3, 4] , UpperCAmelCase_=1 , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = image_size lowerCAmelCase = num_channels lowerCAmelCase = embeddings_size lowerCAmelCase = hidden_sizes lowerCAmelCase = depths lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = hidden_act lowerCAmelCase = num_labels lowerCAmelCase = scope lowerCAmelCase = len(UpperCAmelCase_ ) lowerCAmelCase = out_features lowerCAmelCase = out_indices lowerCAmelCase = num_groups def __snake_case ( self ): lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase = self.get_config() return config, pixel_values, labels def __snake_case ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = BitModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = self.num_labels lowerCAmelCase = BitForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = BitBackbone(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowerCAmelCase = None lowerCAmelCase = BitBackbone(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = config_and_inputs lowerCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : int =(BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __a : Any =( {"""feature-extraction""": BitModel, """image-classification""": BitForImageClassification} if is_torch_available() else {} ) __a : Tuple =False __a : List[str] =False __a : Dict =False __a : Tuple =False __a : int =False def __snake_case ( self ): lowerCAmelCase = BitModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ ) def __snake_case ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def __snake_case ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def __snake_case ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def __snake_case ( self ): pass def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(UpperCAmelCase_ ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(config=UpperCAmelCase_ ) for name, module in model.named_modules(): if isinstance(UpperCAmelCase_ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def __snake_case ( self ): def check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() with torch.no_grad(): lowerCAmelCase = model(**self._prepare_for_class(UpperCAmelCase_ , UpperCAmelCase_ ) ) lowerCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase_ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: lowerCAmelCase = layer_type lowerCAmelCase = True check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase = True check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def __snake_case ( self ): pass def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase_ ) @slow def __snake_case ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = BitModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCAmelCase ( ): lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __snake_case ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self ): lowerCAmelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(UpperCAmelCase_ ) lowerCAmelCase = self.default_image_processor lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**UpperCAmelCase_ ) # verify the logits lowerCAmelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([[-0.6526, -0.5263, -1.4398]] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @require_torch class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Optional[int] =(BitBackbone,) if is_torch_available() else () __a : Dict =BitConfig __a : Tuple =False def __snake_case ( self ): lowerCAmelCase = BitModelTester(self )
710
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ ={ """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """asapp/sew-tiny-100k""": """https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json""", # See all SEW models at https://huggingface.co/models?filter=sew } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] ="""sew""" def __init__( self , UpperCAmelCase_=32 , UpperCAmelCase_=7_68 , UpperCAmelCase_=12 , UpperCAmelCase_=12 , UpperCAmelCase_=30_72 , UpperCAmelCase_=2 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-5 , UpperCAmelCase_="group" , UpperCAmelCase_="gelu" , UpperCAmelCase_=(64, 1_28, 1_28, 1_28, 1_28, 2_56, 2_56, 2_56, 2_56, 5_12, 5_12, 5_12, 5_12) , UpperCAmelCase_=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , UpperCAmelCase_=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , UpperCAmelCase_=False , UpperCAmelCase_=1_28 , UpperCAmelCase_=16 , UpperCAmelCase_=True , UpperCAmelCase_=0.05 , UpperCAmelCase_=10 , UpperCAmelCase_=2 , UpperCAmelCase_=0.0 , UpperCAmelCase_=10 , UpperCAmelCase_=0 , UpperCAmelCase_="mean" , UpperCAmelCase_=False , UpperCAmelCase_=False , UpperCAmelCase_=2_56 , UpperCAmelCase_=0 , UpperCAmelCase_=1 , UpperCAmelCase_=2 , **UpperCAmelCase_ , ): super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ ) lowerCAmelCase = hidden_size lowerCAmelCase = feat_extract_norm lowerCAmelCase = feat_extract_activation lowerCAmelCase = list(UpperCAmelCase_ ) lowerCAmelCase = list(UpperCAmelCase_ ) lowerCAmelCase = list(UpperCAmelCase_ ) lowerCAmelCase = conv_bias lowerCAmelCase = num_conv_pos_embeddings lowerCAmelCase = num_conv_pos_embedding_groups lowerCAmelCase = len(self.conv_dim ) lowerCAmelCase = num_hidden_layers lowerCAmelCase = intermediate_size lowerCAmelCase = squeeze_factor lowerCAmelCase = hidden_act lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_dropout lowerCAmelCase = attention_dropout lowerCAmelCase = activation_dropout lowerCAmelCase = feat_proj_dropout lowerCAmelCase = final_dropout lowerCAmelCase = layerdrop lowerCAmelCase = layer_norm_eps lowerCAmelCase = initializer_range lowerCAmelCase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" F"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCAmelCase = apply_spec_augment lowerCAmelCase = mask_time_prob lowerCAmelCase = mask_time_length lowerCAmelCase = mask_time_min_masks lowerCAmelCase = mask_feature_prob lowerCAmelCase = mask_feature_length lowerCAmelCase = mask_feature_min_masks # ctc loss lowerCAmelCase = ctc_loss_reduction lowerCAmelCase = ctc_zero_infinity # sequence classification lowerCAmelCase = use_weighted_layer_sum lowerCAmelCase = classifier_proj_size @property def __snake_case ( self ): return functools.reduce(operator.mul , self.conv_stride , 1 )
711
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline UpperCAmelCase_ =datasets.utils.logging.get_logger(__name__) @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __a : Optional[datasets.Features] =None __a : str ="utf-8" __a : Optional[str] =None __a : Optional[str] =None __a : bool =True # deprecated __a : Optional[int] =None # deprecated __a : int =1_0 << 2_0 # 10MB __a : Optional[bool] =None class __UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __a : str =JsonConfig def __snake_case ( self ): if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowerCAmelCase = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def __snake_case ( self , UpperCAmelCase_ ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase_ , (str, list, tuple) ): lowerCAmelCase = data_files if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase_ , gen_kwargs={'''files''': files} ) ) return splits def __snake_case ( self , UpperCAmelCase_ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCAmelCase = self.config.features.arrow_schema.field(UpperCAmelCase_ ).type lowerCAmelCase = pa_table.append_column(UpperCAmelCase_ , pa.array([None] * len(UpperCAmelCase_ ) , type=UpperCAmelCase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase = table_cast(UpperCAmelCase_ , self.config.features.arrow_schema ) return pa_table def __snake_case ( self , UpperCAmelCase_ ): for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) # We keep only the field we are interested in lowerCAmelCase = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(UpperCAmelCase_ , (list, tuple) ): lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} else: lowerCAmelCase = dataset lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) yield file_idx, self._cast_table(UpperCAmelCase_ ) # If the file has one json object per line else: with open(UpperCAmelCase_ , '''rb''' ) as f: lowerCAmelCase = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCAmelCase = max(self.config.chunksize // 32 , 16 << 10 ) lowerCAmelCase = ( self.config.encoding_errors if self.config.encoding_errors is not None else '''strict''' ) while True: lowerCAmelCase = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(UpperCAmelCase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCAmelCase = batch.decode(self.config.encoding , errors=UpperCAmelCase_ ).encode('''utf-8''' ) try: while True: try: lowerCAmelCase = paj.read_json( io.BytesIO(UpperCAmelCase_ ) , read_options=paj.ReadOptions(block_size=UpperCAmelCase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(UpperCAmelCase_ , pa.ArrowInvalid ) and "straddling" not in str(UpperCAmelCase_ ) or block_size > len(UpperCAmelCase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(UpperCAmelCase_ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # list is the only sequence type supported in JSON try: lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(UpperCAmelCase_ ) break else: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase_ ) batch_idx += 1
33
0
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCAmelCase_ ="""src/diffusers""" UpperCAmelCase_ =""".""" # This is to make sure the diffusers module imported is the one in the repo. UpperCAmelCase_ =importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) UpperCAmelCase_ =spec.loader.load_module() def UpperCAmelCase ( _snake_case , _snake_case ): return line.startswith(_snake_case ) or len(_snake_case ) <= 1 or re.search(R'''^\s*\)(\s*->.*:|:)\s*$''' , _snake_case ) is not None def UpperCAmelCase ( _snake_case ): lowerCAmelCase = object_name.split('''.''' ) lowerCAmelCase = 0 # First let's find the module where our object lives. lowerCAmelCase = parts[i] while i < len(_snake_case ) and not os.path.isfile(os.path.join(_snake_case , F"""{module}.py""" ) ): i += 1 if i < len(_snake_case ): lowerCAmelCase = os.path.join(_snake_case , parts[i] ) if i >= len(_snake_case ): raise ValueError(F"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(_snake_case , F"""{module}.py""" ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase = f.readlines() # Now let's find the class / func in the code! lowerCAmelCase = '''''' lowerCAmelCase = 0 for name in parts[i + 1 :]: while ( line_index < len(_snake_case ) and re.search(RF"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(_snake_case ): raise ValueError(F""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCAmelCase = line_index while line_index < len(_snake_case ) and _should_continue(lines[line_index] , _snake_case ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCAmelCase = lines[start_index:line_index] return "".join(_snake_case ) UpperCAmelCase_ =re.compile(R"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") UpperCAmelCase_ =re.compile(R"""^\s*(\S+)->(\S+)(\s+.*|$)""") UpperCAmelCase_ =re.compile(R"""<FILL\s+[^>]*>""") def UpperCAmelCase ( _snake_case ): lowerCAmelCase = code.split('''\n''' ) lowerCAmelCase = 0 while idx < len(_snake_case ) and len(lines[idx] ) == 0: idx += 1 if idx < len(_snake_case ): return re.search(R'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def UpperCAmelCase ( _snake_case ): lowerCAmelCase = len(get_indent(_snake_case ) ) > 0 if has_indent: lowerCAmelCase = F"""class Bla:\n{code}""" lowerCAmelCase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=_snake_case ) lowerCAmelCase = black.format_str(_snake_case , mode=_snake_case ) lowerCAmelCase , lowerCAmelCase = style_docstrings_in_code(_snake_case ) return result[len('''class Bla:\n''' ) :] if has_indent else result def UpperCAmelCase ( _snake_case , _snake_case=False ): with open(_snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase = f.readlines() lowerCAmelCase = [] lowerCAmelCase = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(_snake_case ): lowerCAmelCase = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = search.groups() lowerCAmelCase = find_code_in_diffusers(_snake_case ) lowerCAmelCase = get_indent(_snake_case ) lowerCAmelCase = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCAmelCase = theoretical_indent lowerCAmelCase = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCAmelCase = True while line_index < len(_snake_case ) and should_continue: line_index += 1 if line_index >= len(_snake_case ): break lowerCAmelCase = lines[line_index] lowerCAmelCase = _should_continue(_snake_case , _snake_case ) and re.search(F"""^{indent}# End copy""" , _snake_case ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCAmelCase = lines[start_index:line_index] lowerCAmelCase = ''''''.join(_snake_case ) # Remove any nested `Copied from` comments to avoid circular copies lowerCAmelCase = [line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(_snake_case ) is None] lowerCAmelCase = '''\n'''.join(_snake_case ) # Before comparing, use the `replace_pattern` on the original code. if len(_snake_case ) > 0: lowerCAmelCase = replace_pattern.replace('''with''' , '''''' ).split(''',''' ) lowerCAmelCase = [_re_replace_pattern.search(_snake_case ) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = pattern.groups() lowerCAmelCase = re.sub(_snake_case , _snake_case , _snake_case ) if option.strip() == "all-casing": lowerCAmelCase = re.sub(obja.lower() , obja.lower() , _snake_case ) lowerCAmelCase = re.sub(obja.upper() , obja.upper() , _snake_case ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCAmelCase = blackify(lines[start_index - 1] + theoretical_code ) lowerCAmelCase = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowerCAmelCase = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCAmelCase = start_index + 1 if overwrite and len(_snake_case ) > 0: # Warn the user a file has been modified. print(F"""Detected changes, rewriting {filename}.""" ) with open(_snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_snake_case ) return diffs def UpperCAmelCase ( _snake_case = False ): lowerCAmelCase = glob.glob(os.path.join(_snake_case , '''**/*.py''' ) , recursive=_snake_case ) lowerCAmelCase = [] for filename in all_files: lowerCAmelCase = is_copy_consistent(_snake_case , _snake_case ) diffs += [F"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(_snake_case ) > 0: lowerCAmelCase = '''\n'''.join(_snake_case ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") UpperCAmelCase_ =parser.parse_args() check_copies(args.fix_and_overwrite)
712
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] ="""maskformer-swin""" __a : Optional[int] ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , UpperCAmelCase_=2_24 , UpperCAmelCase_=4 , UpperCAmelCase_=3 , UpperCAmelCase_=96 , UpperCAmelCase_=[2, 2, 6, 2] , UpperCAmelCase_=[3, 6, 12, 24] , UpperCAmelCase_=7 , UpperCAmelCase_=4.0 , UpperCAmelCase_=True , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.1 , UpperCAmelCase_="gelu" , UpperCAmelCase_=False , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-5 , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = embed_dim lowerCAmelCase = depths lowerCAmelCase = len(UpperCAmelCase_ ) lowerCAmelCase = num_heads lowerCAmelCase = window_size lowerCAmelCase = mlp_ratio lowerCAmelCase = qkv_bias lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = drop_path_rate lowerCAmelCase = hidden_act lowerCAmelCase = use_absolute_embeddings lowerCAmelCase = layer_norm_eps lowerCAmelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase = int(embed_dim * 2 ** (len(UpperCAmelCase_ ) - 1) ) lowerCAmelCase = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase_ ) + 1 )] lowerCAmelCase , lowerCAmelCase = get_aligned_output_features_output_indices( out_features=UpperCAmelCase_ , out_indices=UpperCAmelCase_ , stage_names=self.stage_names )
33
0
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =CLIPTokenizer __a : Dict =CLIPTokenizerFast __a : Optional[int] =True __a : int ={} __a : Any =False def __snake_case ( self ): super().setUp() # fmt: off lowerCAmelCase = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>'''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self , **UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def __snake_case ( self , **UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''lower newer''' lowerCAmelCase = '''lower newer''' return input_text, output_text def __snake_case ( self ): lowerCAmelCase = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCAmelCase = '''lower newer''' lowerCAmelCase = ['''lo''', '''w''', '''er</w>''', '''n''', '''e''', '''w''', '''er</w>'''] lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = tokens + [tokenizer.unk_token] lowerCAmelCase = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , UpperCAmelCase_ ) @require_ftfy def __snake_case ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = self.tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) lowerCAmelCase = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) lowerCAmelCase = '''A\n\'ll 11p223RF☆ho!!to?\'d\'d\'\'d of a cat to-$\'\'d.''' lowerCAmelCase = tokenizer_s.tokenize(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways lowerCAmelCase = '''xa\u0303y''' + ''' ''' + '''x\xe3y''' lowerCAmelCase = tokenizer_s.tokenize(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Test that the tokenization is identical on unicode of space type lowerCAmelCase = [ '''\u0009''', # (horizontal tab, '\t') '''\u000B''', # (vertical tab) '''\u000C''', # (form feed) '''\u0020''', # (space, ' ') '''\u200E''', # (left-to-right mark):w '''\u200F''', # (right-to-left mark) ] for unicode_seq in spaces_unicodes: lowerCAmelCase = tokenizer_s.tokenize(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Test that the tokenization is identical on unicode of line break type lowerCAmelCase = [ '''\u000A''', # (line feed, '\n') '''\r\n''', # (carriage return and line feed, '\r\n') '''\u000D''', # (carriage return, '\r') '''\r''', # (carriage return, '\r') '''\u000D''', # (carriage return, '\r') '''\u2028''', # (line separator) '''\u2029''', # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: lowerCAmelCase = tokenizer_s.tokenize(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_r.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` lowerCAmelCase = F"""{text_of_1_token} {text_of_1_token}""" lowerCAmelCase = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase_ , use_fast=UpperCAmelCase_ , ) lowerCAmelCase = tokenizer_r(UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCAmelCase_ ) + 1, len(UpperCAmelCase_ ) + 1 + len(UpperCAmelCase_ )) , ) lowerCAmelCase = F""" {text}""" lowerCAmelCase = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase_ , use_fast=UpperCAmelCase_ , ) lowerCAmelCase = tokenizer_r(UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(UpperCAmelCase_ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCAmelCase_ ) + 1, 1 + len(UpperCAmelCase_ ) + 1 + len(UpperCAmelCase_ )) , ) def __snake_case ( self ): # Test related to the breaking change introduced in transformers v4.17.0 # We need to check that an error in raised when the user try to load a previous version of the tokenizer. with self.assertRaises(UpperCAmelCase_ ) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''' ) self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''' ) ) @require_ftfy def __snake_case ( self ): super().test_tokenization_python_rust_equals() def __snake_case ( self ): # CLIP always lower cases letters pass
713
from collections.abc import Sequence def UpperCAmelCase ( _snake_case , _snake_case = False ): if not arr: return 0 lowerCAmelCase = 0 if allow_empty_subarrays else float('''-inf''' ) lowerCAmelCase = 0.0 for num in arr: lowerCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) lowerCAmelCase = max(_snake_case , _snake_case ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase_ =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'''{max_subarray_sum(nums) = }''')
33
0
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def UpperCAmelCase ( _snake_case ): return 1.0 / (1.0 + np.exp(-_outputs )) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = np.max(_outputs , axis=-1 , keepdims=_snake_case ) lowerCAmelCase = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Optional[int] ="""sigmoid""" __a : List[Any] ="""softmax""" __a : Dict ="""none""" @add_end_docstrings( __UpperCAmelCase , r""" return_all_scores (`bool`, *optional*, defaults to `False`): Whether to return all prediction scores or just the one of the predicted class. function_to_apply (`str`, *optional*, defaults to `\"default\"`): The function to apply to the model outputs in order to retrieve the scores. Accepts four different values: - `\"default\"`: if the model has a single label, will apply the sigmoid function on the output. If the model has several labels, will apply the softmax function on the output. - `\"sigmoid\"`: Applies the sigmoid function on the output. - `\"softmax\"`: Applies the softmax function on the output. - `\"none\"`: Does not apply any function on the output. """ , ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] =False __a : Any =ClassificationFunction.NONE def __init__( self , **UpperCAmelCase_ ): super().__init__(**UpperCAmelCase_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def __snake_case ( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_="" , **UpperCAmelCase_ ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" lowerCAmelCase = tokenizer_kwargs lowerCAmelCase = {} if hasattr(self.model.config , '''return_all_scores''' ) and return_all_scores is None: lowerCAmelCase = self.model.config.return_all_scores if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or top_k is None: lowerCAmelCase = top_k lowerCAmelCase = False elif return_all_scores is not None: warnings.warn( '''`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of''' ''' `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.''' , UpperCAmelCase_ , ) if return_all_scores: lowerCAmelCase = None else: lowerCAmelCase = 1 if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: lowerCAmelCase = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): lowerCAmelCase = super().__call__(*UpperCAmelCase_ , **UpperCAmelCase_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. lowerCAmelCase = '''top_k''' not in kwargs if isinstance(args[0] , UpperCAmelCase_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def __snake_case ( self , UpperCAmelCase_ , **UpperCAmelCase_ ): lowerCAmelCase = self.framework if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): return self.tokenizer(**UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and len(UpperCAmelCase_ ) == 1 and isinstance(inputs[0] , UpperCAmelCase_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( '''The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a''' ''' dictionary `{"text": "My text", "text_pair": "My pair"}` in order to send a text pair.''' ) return self.tokenizer(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): return self.model(**UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=None , UpperCAmelCase_=1 , UpperCAmelCase_=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: lowerCAmelCase = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: lowerCAmelCase = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , '''function_to_apply''' ) and function_to_apply is None: lowerCAmelCase = self.model.config.function_to_apply else: lowerCAmelCase = ClassificationFunction.NONE lowerCAmelCase = model_outputs['''logits'''][0] lowerCAmelCase = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: lowerCAmelCase = sigmoid(UpperCAmelCase_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: lowerCAmelCase = softmax(UpperCAmelCase_ ) elif function_to_apply == ClassificationFunction.NONE: lowerCAmelCase = outputs else: raise ValueError(F"""Unrecognized `function_to_apply` argument: {function_to_apply}""" ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} lowerCAmelCase = [ {'''label''': self.model.config.idalabel[i], '''score''': score.item()} for i, score in enumerate(UpperCAmelCase_ ) ] if not _legacy: dict_scores.sort(key=lambda UpperCAmelCase_ : x["score"] , reverse=UpperCAmelCase_ ) if top_k is not None: lowerCAmelCase = dict_scores[:top_k] return dict_scores
714
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =BertJapaneseTokenizer __a : Optional[int] =False __a : int =True def __snake_case ( self ): super().setUp() lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(do_lower_case=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer( do_lower_case=UpperCAmelCase_ , normalize_text=UpperCAmelCase_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(normalize_text=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(do_lower_case=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(normalize_text=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(trim_whitespace=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(do_lower_case=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(normalize_text=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(trim_whitespace=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = WordpieceTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) lowerCAmelCase = tokenizer.subword_tokenizer lowerCAmelCase = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) lowerCAmelCase = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =BertJapaneseTokenizer __a : Optional[int] =False def __snake_case ( self ): super().setUp() lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , **UpperCAmelCase_ ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( UpperCAmelCase_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = CharacterTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) lowerCAmelCase = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
33
0
import requests from bsa import BeautifulSoup def UpperCAmelCase ( _snake_case , _snake_case ): '''simple docstring''' lowerCAmelCase = BeautifulSoup(requests.get(_snake_case , params=_snake_case ).content , '''html.parser''' ) lowerCAmelCase = soup.find('''div''' , attrs={'''class''': '''gs_ri'''} ) lowerCAmelCase = div.find('''div''' , attrs={'''class''': '''gs_fl'''} ).find_all('''a''' ) return anchors[2].get_text() if __name__ == "__main__": UpperCAmelCase_ ={ """title""": ( """Precisely geometry controlled microsupercapacitors for ultrahigh areal """ """capacitance, volumetric capacitance, and energy density""" ), """journal""": """Chem. Mater.""", """volume""": 30, """pages""": """3979-3990""", """year""": 2018, """hl""": """en""", } print(get_citation("""https://scholar.google.com/scholar_lookup""", params=params))
715
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) UpperCAmelCase_ ="""hf-internal-testing/tiny-random-bert""" UpperCAmelCase_ =os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") UpperCAmelCase_ ="""9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCAmelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertTrue(os.path.isfile(UpperCAmelCase_ ) ) # File is cached at the same place the second time. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Using a specific revision to test the full commit hash. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''9b8c223''' ) self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): lowerCAmelCase = cached_file('''tiny-random-bert''' , UpperCAmelCase_ ) with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCAmelCase_ , '''.no_exist''' , UpperCAmelCase_ , '''conf''' ) ) ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , local_files_only=UpperCAmelCase_ , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = mock.Mock() lowerCAmelCase = 5_00 lowerCAmelCase = {} lowerCAmelCase = HTTPError lowerCAmelCase = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCAmelCase_ ) as mock_head: lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) # This check we did call the fake head request mock_head.assert_called() def __snake_case ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) def __snake_case ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCAmelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ , revision='''ahaha''' ) lowerCAmelCase = get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCAmelCase = json.loads(open(UpperCAmelCase_ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_68 ) def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase = Path(UpperCAmelCase_ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCAmelCase_ , '''a.txt''' ) , str(UpperCAmelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCAmelCase_ , '''b.txt''' ) )
33
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =BertJapaneseTokenizer __a : Optional[int] =False __a : int =True def __snake_case ( self ): super().setUp() lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(do_lower_case=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer( do_lower_case=UpperCAmelCase_ , normalize_text=UpperCAmelCase_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(normalize_text=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(do_lower_case=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(normalize_text=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(trim_whitespace=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(do_lower_case=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(normalize_text=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(trim_whitespace=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = WordpieceTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) lowerCAmelCase = tokenizer.subword_tokenizer lowerCAmelCase = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) lowerCAmelCase = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =BertJapaneseTokenizer __a : Optional[int] =False def __snake_case ( self ): super().setUp() lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , **UpperCAmelCase_ ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( UpperCAmelCase_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = CharacterTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) lowerCAmelCase = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
716
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = "arrow" , **UpperCAmelCase_ , ): super().__init__( split=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , keep_in_memory=UpperCAmelCase_ , streaming=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = load_from_cache_file lowerCAmelCase = file_format lowerCAmelCase = Spark( df=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , working_dir=UpperCAmelCase_ , **UpperCAmelCase_ , ) def __snake_case ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
33
0
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) UpperCAmelCase_ ={ """iou_prediction_head.layers.0""": """iou_prediction_head.proj_in""", """iou_prediction_head.layers.1""": """iou_prediction_head.layers.0""", """iou_prediction_head.layers.2""": """iou_prediction_head.proj_out""", """mask_decoder.output_upscaling.0""": """mask_decoder.upscale_conv1""", """mask_decoder.output_upscaling.1""": """mask_decoder.upscale_layer_norm""", """mask_decoder.output_upscaling.3""": """mask_decoder.upscale_conv2""", """mask_downscaling.0""": """mask_embed.conv1""", """mask_downscaling.1""": """mask_embed.layer_norm1""", """mask_downscaling.3""": """mask_embed.conv2""", """mask_downscaling.4""": """mask_embed.layer_norm2""", """mask_downscaling.6""": """mask_embed.conv3""", """point_embeddings""": """point_embed""", """pe_layer.positional_encoding_gaussian_matrix""": """shared_embedding.positional_embedding""", """image_encoder""": """vision_encoder""", """neck.0""": """neck.conv1""", """neck.1""": """neck.layer_norm1""", """neck.2""": """neck.conv2""", """neck.3""": """neck.layer_norm2""", """patch_embed.proj""": """patch_embed.projection""", """.norm""": """.layer_norm""", """blocks""": """layers""", } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = {} state_dict.pop('''pixel_mean''' , _snake_case ) state_dict.pop('''pixel_std''' , _snake_case ) lowerCAmelCase = R'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): lowerCAmelCase = int(re.match(_snake_case , _snake_case ).group(2 ) ) if layer_nb == 0: lowerCAmelCase = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: lowerCAmelCase = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: lowerCAmelCase = key.replace('''layers.2''' , '''proj_out''' ) lowerCAmelCase = value lowerCAmelCase = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case="ybelkada/segment-anything" ): lowerCAmelCase = hf_hub_download(_snake_case , F"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: lowerCAmelCase = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase = SamConfig( vision_config=_snake_case , ) elif "sam_vit_h" in model_name: lowerCAmelCase = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase = SamConfig( vision_config=_snake_case , ) lowerCAmelCase = torch.load(_snake_case , map_location='''cpu''' ) lowerCAmelCase = replace_keys(_snake_case ) lowerCAmelCase = SamImageProcessor() lowerCAmelCase = SamProcessor(image_processor=_snake_case ) lowerCAmelCase = SamModel(_snake_case ) hf_model.load_state_dict(_snake_case ) lowerCAmelCase = hf_model.to('''cuda''' ) lowerCAmelCase = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' lowerCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert('''RGB''' ) lowerCAmelCase = [[[400, 650]]] lowerCAmelCase = [[1]] lowerCAmelCase = processor(images=np.array(_snake_case ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**_snake_case ) lowerCAmelCase = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 lowerCAmelCase = processor( images=np.array(_snake_case ) , input_points=_snake_case , input_labels=_snake_case , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**_snake_case ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 lowerCAmelCase = ((75, 275, 1725, 850),) lowerCAmelCase = processor(images=np.array(_snake_case ) , input_boxes=_snake_case , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**_snake_case ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. lowerCAmelCase = [[[400, 650], [800, 650]]] lowerCAmelCase = [[1, 1]] lowerCAmelCase = processor( images=np.array(_snake_case ) , input_points=_snake_case , input_labels=_snake_case , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): lowerCAmelCase = hf_model(**_snake_case ) lowerCAmelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() UpperCAmelCase_ =["""sam_vit_b_01ec64""", """sam_vit_h_4b8939""", """sam_vit_l_0b3195"""] parser.add_argument( """--model_name""", default="""sam_vit_h_4b8939""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) parser.add_argument( """--model_hub_id""", default="""ybelkada/segment-anything""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) UpperCAmelCase_ =parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
717
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase ( _snake_case = 3 ): if isinstance(_snake_case , _snake_case ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(_snake_case ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) lowerCAmelCase = QuantumRegister(_snake_case , '''qr''' ) lowerCAmelCase = ClassicalRegister(_snake_case , '''cr''' ) lowerCAmelCase = QuantumCircuit(_snake_case , _snake_case ) lowerCAmelCase = number_of_qubits for i in range(_snake_case ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(_snake_case ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , _snake_case , _snake_case ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(_snake_case , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(_snake_case , _snake_case ) # simulate with 10000 shots lowerCAmelCase = Aer.get_backend('''qasm_simulator''' ) lowerCAmelCase = execute(_snake_case , _snake_case , shots=10000 ) return job.result().get_counts(_snake_case ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
33
0
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = "arrow" , **UpperCAmelCase_ , ): super().__init__( split=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , keep_in_memory=UpperCAmelCase_ , streaming=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = load_from_cache_file lowerCAmelCase = file_format lowerCAmelCase = Spark( df=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , working_dir=UpperCAmelCase_ , **UpperCAmelCase_ , ) def __snake_case ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
718
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Any =1 @register_to_config def __init__( self , UpperCAmelCase_=20_00 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=1E-3 ): lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = torch.linspace(1 , self.config.sampling_eps , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCAmelCase = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCAmelCase = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCAmelCase = std.flatten() while len(std.shape ) < len(score.shape ): lowerCAmelCase = std.unsqueeze(-1 ) lowerCAmelCase = -score / std # compute lowerCAmelCase = -1.0 / len(self.timesteps ) lowerCAmelCase = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCAmelCase = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCAmelCase = beta_t.unsqueeze(-1 ) lowerCAmelCase = -0.5 * beta_t * x lowerCAmelCase = torch.sqrt(UpperCAmelCase_ ) lowerCAmelCase = drift - diffusion**2 * score lowerCAmelCase = x + drift * dt # add noise lowerCAmelCase = randn_tensor(x.shape , layout=x.layout , generator=UpperCAmelCase_ , device=x.device , dtype=x.dtype ) lowerCAmelCase = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ): return self.config.num_train_timesteps
33
0
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : int =["""image_processor""", """tokenizer"""] __a : Any ="""BlipImageProcessor""" __a : Union[str, Any] ="""AutoTokenizer""" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = False super().__init__(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = self.image_processor def __call__( self , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = 0 , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = False , UpperCAmelCase_ = False , UpperCAmelCase_ = False , UpperCAmelCase_ = False , UpperCAmelCase_ = True , UpperCAmelCase_ = None , **UpperCAmelCase_ , ): if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None: lowerCAmelCase = self.tokenizer lowerCAmelCase = self.tokenizer( text=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , stride=UpperCAmelCase_ , pad_to_multiple_of=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , return_overflowing_tokens=UpperCAmelCase_ , return_special_tokens_mask=UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ , return_length=UpperCAmelCase_ , verbose=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ , ) return text_encoding # add pixel_values lowerCAmelCase = self.image_processor(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ ) if text is not None: lowerCAmelCase = self.tokenizer( text=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , stride=UpperCAmelCase_ , pad_to_multiple_of=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , return_overflowing_tokens=UpperCAmelCase_ , return_special_tokens_mask=UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ , return_length=UpperCAmelCase_ , verbose=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ , ) else: lowerCAmelCase = None if text_encoding is not None: encoding_image_processor.update(UpperCAmelCase_ ) return encoding_image_processor def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.batch_decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __snake_case ( self ): lowerCAmelCase = self.tokenizer.model_input_names lowerCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
719
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
33
0
import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, AutoModelForMaskedLM, AutoTokenizer, DataCollatorForWholeWordMask, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process UpperCAmelCase_ =logging.getLogger(__name__) UpperCAmelCase_ =list(MODEL_FOR_MASKED_LM_MAPPING.keys()) UpperCAmelCase_ =tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __UpperCamelCase : '''simple docstring''' __a : Optional[str] =field( default=__UpperCAmelCase , metadata={ """help""": ( """The model checkpoint for weights initialization.Don't set if you want to train a model from scratch.""" ) } , ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(__UpperCAmelCase )} , ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) } , ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __a : bool =field( default=__UpperCAmelCase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __a : str =field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __a : bool =field( default=__UpperCAmelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) def __snake_case ( self ): if self.config_overrides is not None and (self.config_name is not None or self.model_name_or_path is not None): raise ValueError( '''--config_overrides can\'t be used in combination with --config_name or --model_name_or_path''' ) @dataclass class __UpperCamelCase : '''simple docstring''' __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) __a : Optional[str] =field(default=__UpperCAmelCase , metadata={"""help""": """The input training data file (a text file)."""} ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """An optional input train ref data file for whole word masking in Chinese."""} , ) __a : Optional[str] =field( default=__UpperCAmelCase , metadata={"""help""": """An optional input validation ref data file for whole word masking in Chinese."""} , ) __a : bool =field( default=__UpperCAmelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __a : Optional[int] =field( default=5 , metadata={ """help""": """The percentage of the train set used as validation set in case there's no validation split""" } , ) __a : Optional[int] =field( default=__UpperCAmelCase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated. Default to the max input length of the model.""" ) } , ) __a : Optional[int] =field( default=__UpperCAmelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __a : float =field( default=0.15 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) __a : bool =field( default=__UpperCAmelCase , metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) } , ) def __snake_case ( self ): if self.train_file is not None: lowerCAmelCase = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json", "txt"], "`train_file` should be a csv, a json or a txt file." if self.validation_file is not None: lowerCAmelCase = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json", "txt"], "`validation_file` should be a csv, a json or a txt file." def UpperCAmelCase ( _snake_case , _snake_case ): with open(_snake_case , '''r''' , encoding='''utf-8''' ) as f: lowerCAmelCase = [json.loads(_snake_case ) for line in f.read().splitlines() if (len(_snake_case ) > 0 and not line.isspace())] assert len(_snake_case ) == len(_snake_case ) lowerCAmelCase = {c: dataset[c] for c in dataset.column_names} lowerCAmelCase = refs return Dataset.from_dict(_snake_case ) def UpperCAmelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = parser.parse_args_into_dataclasses() # Detecting last checkpoint. lowerCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , _snake_case ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. lowerCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name ) if "validation" not in datasets.keys(): lowerCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F"""train[:{data_args.validation_split_percentage}%]""" , ) lowerCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F"""train[{data_args.validation_split_percentage}%:]""" , ) else: lowerCAmelCase = {} if data_args.train_file is not None: lowerCAmelCase = data_args.train_file if data_args.validation_file is not None: lowerCAmelCase = data_args.validation_file lowerCAmelCase = data_args.train_file.split('''.''' )[-1] if extension == "txt": lowerCAmelCase = '''text''' lowerCAmelCase = load_dataset(_snake_case , data_files=_snake_case ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase = { '''cache_dir''': model_args.cache_dir, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.config_name: lowerCAmelCase = AutoConfig.from_pretrained(model_args.config_name , **_snake_case ) elif model_args.model_name_or_path: lowerCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: lowerCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(F"""Overriding config: {model_args.config_overrides}""" ) config.update_from_string(model_args.config_overrides ) logger.info(F"""New config: {config}""" ) lowerCAmelCase = { '''cache_dir''': model_args.cache_dir, '''use_fast''': model_args.use_fast_tokenizer, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.tokenizer_name: lowerCAmelCase = AutoTokenizer.from_pretrained(model_args.tokenizer_name , **_snake_case ) elif model_args.model_name_or_path: lowerCAmelCase = AutoTokenizer.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported by this script.''' '''You can do it from another script, save it, and load it from here, using --tokenizer_name.''' ) if model_args.model_name_or_path: lowerCAmelCase = AutoModelForMaskedLM.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('''Training new model from scratch''' ) lowerCAmelCase = AutoModelForMaskedLM.from_config(_snake_case ) model.resize_token_embeddings(len(_snake_case ) ) # Preprocessing the datasets. # First we tokenize all the texts. if training_args.do_train: lowerCAmelCase = datasets['''train'''].column_names else: lowerCAmelCase = datasets['''validation'''].column_names lowerCAmelCase = '''text''' if '''text''' in column_names else column_names[0] lowerCAmelCase = '''max_length''' if data_args.pad_to_max_length else False def tokenize_function(_snake_case ): # Remove empty lines lowerCAmelCase = [line for line in examples['''text'''] if len(_snake_case ) > 0 and not line.isspace()] return tokenizer(examples['''text'''] , padding=_snake_case , truncation=_snake_case , max_length=data_args.max_seq_length ) lowerCAmelCase = datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=[text_column_name] , load_from_cache_file=not data_args.overwrite_cache , ) # Add the chinese references if provided if data_args.train_ref_file is not None: lowerCAmelCase = add_chinese_references(tokenized_datasets['''train'''] , data_args.train_ref_file ) if data_args.validation_ref_file is not None: lowerCAmelCase = add_chinese_references( tokenized_datasets['''validation'''] , data_args.validation_ref_file ) # If we have ref files, need to avoid it removed by trainer lowerCAmelCase = data_args.train_ref_file or data_args.validation_ref_file if has_ref: lowerCAmelCase = False # Data collator # This one will take care of randomly masking the tokens. lowerCAmelCase = DataCollatorForWholeWordMask(tokenizer=_snake_case , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowerCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=tokenized_datasets['''train'''] if training_args.do_train else None , eval_dataset=tokenized_datasets['''validation'''] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: if last_checkpoint is not None: lowerCAmelCase = last_checkpoint elif model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ): lowerCAmelCase = model_args.model_name_or_path else: lowerCAmelCase = None lowerCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCAmelCase = os.path.join(training_args.output_dir , '''train_results.txt''' ) if trainer.is_world_process_zero(): with open(_snake_case , '''w''' ) as writer: logger.info('''***** Train results *****''' ) for key, value in sorted(train_result.metrics.items() ): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir , '''trainer_state.json''' ) ) # Evaluation lowerCAmelCase = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCAmelCase = trainer.evaluate() lowerCAmelCase = math.exp(eval_output['''eval_loss'''] ) lowerCAmelCase = perplexity lowerCAmelCase = os.path.join(training_args.output_dir , '''eval_results_mlm_wwm.txt''' ) if trainer.is_world_process_zero(): with open(_snake_case , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in sorted(results.items() ): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) return results def UpperCAmelCase ( _snake_case ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
720
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : str =MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = VideoClassificationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ , top_k=2 ) lowerCAmelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): for example in examples: lowerCAmelCase = video_classifier(UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, ] , ) @require_torch def __snake_case ( self ): lowerCAmelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' lowerCAmelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) lowerCAmelCase = pipeline( '''video-classification''' , model=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = video_classifier(UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __snake_case ( self ): pass
33
0
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case ): if height >= 1: move_tower(height - 1 , _snake_case , _snake_case , _snake_case ) move_disk(_snake_case , _snake_case ) move_tower(height - 1 , _snake_case , _snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case ): print('''moving disk from''' , _snake_case , '''to''' , _snake_case ) def UpperCAmelCase ( ): lowerCAmelCase = int(input('''Height of hanoi: ''' ).strip() ) move_tower(_snake_case , '''A''' , '''B''' , '''C''' ) if __name__ == "__main__": main()
721
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any ="""hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __snake_case ( self , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCAmelCase_ ) ) lowerCAmelCase = np.random.RandomState(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # warmup pass to apply optimizations lowerCAmelCase = pipe(**self.get_dummy_inputs() ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @property def __snake_case ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __snake_case ( self ): lowerCAmelCase = ort.SessionOptions() lowerCAmelCase = False return options def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCAmelCase_ , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
33
0
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =["""image_processor""", """tokenizer"""] __a : List[str] ="""OwlViTImageProcessor""" __a : List[Any] =("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ ): lowerCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase_ , ) lowerCAmelCase = kwargs.pop('''feature_extractor''' ) lowerCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase_ , UpperCAmelCase_ ) def __call__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_="max_length" , UpperCAmelCase_="np" , **UpperCAmelCase_ ): if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''' ) if text is not None: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or (isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and not isinstance(text[0] , UpperCAmelCase_ )): lowerCAmelCase = [self.tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ )] elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and isinstance(text[0] , UpperCAmelCase_ ): lowerCAmelCase = [] # Maximum number of queries across batch lowerCAmelCase = max([len(UpperCAmelCase_ ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(UpperCAmelCase_ ) != max_num_queries: lowerCAmelCase = t + [''' '''] * (max_num_queries - len(UpperCAmelCase_ )) lowerCAmelCase = self.tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) encodings.append(UpperCAmelCase_ ) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''' ) if return_tensors == "np": lowerCAmelCase = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) lowerCAmelCase = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp lowerCAmelCase = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) lowerCAmelCase = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch lowerCAmelCase = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0 ) lowerCAmelCase = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf lowerCAmelCase = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) lowerCAmelCase = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) else: raise ValueError('''Target return tensor type could not be returned''' ) lowerCAmelCase = BatchEncoding() lowerCAmelCase = input_ids lowerCAmelCase = attention_mask if query_images is not None: lowerCAmelCase = BatchEncoding() lowerCAmelCase = self.image_processor( UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ).pixel_values lowerCAmelCase = query_pixel_values if images is not None: lowerCAmelCase = self.image_processor(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) if text is not None and images is not None: lowerCAmelCase = image_features.pixel_values return encoding elif query_images is not None and images is not None: lowerCAmelCase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase_ ) , tensor_type=UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.image_processor.post_process(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.image_processor.post_process_object_detection(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.image_processor.post_process_image_guided_detection(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.batch_decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) def __snake_case ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ): return self.tokenizer.decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) @property def __snake_case ( self ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase_ , ) return self.image_processor_class @property def __snake_case ( self ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase_ , ) return self.image_processor
700
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCAmelCase ( _snake_case ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('''/''' ) lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase = torch.load(os.path.join(_snake_case , '''pytorch_model.bin''' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase , lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_snake_case ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F"""bertarized_{os.path.basename(_snake_case )}""" ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(_snake_case , os.path.join(_snake_case , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) UpperCAmelCase_ =parser.parse_args() main(args)
33
0
import argparse import os import re UpperCAmelCase_ ="""src/transformers/models/auto""" # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict UpperCAmelCase_ =re.compile(R"""[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict""") # re pattern that matches identifiers in mappings UpperCAmelCase_ =re.compile(R"""\s*\(\s*\"(\S[^\"]+)\"""") def UpperCAmelCase ( _snake_case , _snake_case = False ): with open(_snake_case , '''r''' , encoding='''utf-8''' ) as f: lowerCAmelCase = f.read() lowerCAmelCase = content.split('''\n''' ) lowerCAmelCase = [] lowerCAmelCase = 0 while line_idx < len(_snake_case ): if _re_intro_mapping.search(lines[line_idx] ) is not None: lowerCAmelCase = len(re.search(R'''^(\s*)\S''' , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(''' ''' * indent + '''(''' ): new_lines.append(lines[line_idx] ) line_idx += 1 lowerCAmelCase = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": lowerCAmelCase = line_idx while not lines[line_idx].startswith(''' ''' * indent + ''')''' ): line_idx += 1 blocks.append('''\n'''.join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers lowerCAmelCase = sorted(_snake_case , key=lambda _snake_case : _re_identifier.search(_snake_case ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(_snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write('''\n'''.join(_snake_case ) ) elif "\n".join(_snake_case ) != content: return True def UpperCAmelCase ( _snake_case = False ): lowerCAmelCase = [os.path.join(_snake_case , _snake_case ) for f in os.listdir(_snake_case ) if f.endswith('''.py''' )] lowerCAmelCase = [sort_auto_mapping(_snake_case , overwrite=_snake_case ) for fname in fnames] if not overwrite and any(_snake_case ): lowerCAmelCase = [f for f, d in zip(_snake_case , _snake_case ) if d] raise ValueError( F"""The following files have auto mappings that need sorting: {", ".join(_snake_case )}. Run `make style` to fix""" ''' this.''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") UpperCAmelCase_ =parser.parse_args() sort_all_auto_mappings(not args.check_only)
701
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } UpperCAmelCase_ ={ """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } UpperCAmelCase_ ={ """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = set() lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase = char lowerCAmelCase = set(_snake_case ) return pairs class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =VOCAB_FILES_NAMES __a : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP __a : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , **UpperCAmelCase_ , ): super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = merges_file lowerCAmelCase = {} lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 lowerCAmelCase = 3 self.add_from_file(UpperCAmelCase_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase_ , encoding='''utf-8''' ) as merges_handle: lowerCAmelCase = merges_handle.read().split('''\n''' )[:-1] lowerCAmelCase = [tuple(merge.split()[:-1] ) for merge in merges] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __snake_case ( self ): return len(self.encoder ) def __snake_case ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self , UpperCAmelCase_ ): if token in self.cache: return self.cache[token] lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowerCAmelCase = get_pairs(UpperCAmelCase_ ) if not pairs: return token while True: lowerCAmelCase = min(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : self.bpe_ranks.get(UpperCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase = bigram lowerCAmelCase = [] lowerCAmelCase = 0 while i < len(UpperCAmelCase_ ): try: lowerCAmelCase = word.index(UpperCAmelCase_ , UpperCAmelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase = j if word[i] == first and i < len(UpperCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = new_word if len(UpperCAmelCase_ ) == 1: break else: lowerCAmelCase = get_pairs(UpperCAmelCase_ ) lowerCAmelCase = '''@@ '''.join(UpperCAmelCase_ ) lowerCAmelCase = word[:-4] lowerCAmelCase = word return word def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [] lowerCAmelCase = re.findall(r'''\S+\n?''' , UpperCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __snake_case ( self , UpperCAmelCase_ ): return self.encoder.get(UpperCAmelCase_ , self.encoder.get(self.unk_token ) ) def __snake_case ( self , UpperCAmelCase_ ): return self.decoder.get(UpperCAmelCase_ , self.unk_token ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = ''' '''.join(UpperCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.merges_file , UpperCAmelCase_ ) return out_vocab_file, out_merge_file def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(UpperCAmelCase_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCAmelCase = f.readlines() for lineTmp in lines: lowerCAmelCase = lineTmp.strip() lowerCAmelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowerCAmelCase = line[:idx] lowerCAmelCase = len(self.encoder )
33
0
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : int =(UniPCMultistepScheduler,) __a : Dict =(("""num_inference_steps""", 2_5),) def __snake_case ( self , **UpperCAmelCase_ ): lowerCAmelCase = { '''num_train_timesteps''': 10_00, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**UpperCAmelCase_ ) return config def __snake_case ( self , UpperCAmelCase_=0 , **UpperCAmelCase_ ): lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('''num_inference_steps''' , UpperCAmelCase_ ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config(**UpperCAmelCase_ ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(UpperCAmelCase_ ) # copy over dummy past residuals lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase_ ) lowerCAmelCase = scheduler_class.from_pretrained(UpperCAmelCase_ ) new_scheduler.set_timesteps(UpperCAmelCase_ ) # copy over dummy past residuals lowerCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase , lowerCAmelCase = sample, sample for t in range(UpperCAmelCase_ , time_step + scheduler.config.solver_order + 1 ): lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample lowerCAmelCase = new_scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def __snake_case ( self , UpperCAmelCase_=0 , **UpperCAmelCase_ ): lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('''num_inference_steps''' , UpperCAmelCase_ ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(UpperCAmelCase_ ) # copy over dummy past residuals (must be after setting timesteps) lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase_ ) lowerCAmelCase = scheduler_class.from_pretrained(UpperCAmelCase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase_ ) # copy over dummy past residual (must be after setting timesteps) lowerCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample lowerCAmelCase = new_scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def __snake_case ( self , UpperCAmelCase_=None , **UpperCAmelCase_ ): if scheduler is None: lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(**UpperCAmelCase_ ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(**UpperCAmelCase_ ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) lowerCAmelCase = 10 lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ).prev_sample return sample def __snake_case ( self ): lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('''num_inference_steps''' , UpperCAmelCase_ ) for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase_ , '''set_timesteps''' ): scheduler.set_timesteps(UpperCAmelCase_ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase_ , '''set_timesteps''' ): lowerCAmelCase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] lowerCAmelCase = scheduler.timesteps[5] lowerCAmelCase = scheduler.timesteps[6] lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __snake_case ( self ): # make sure that iterating over schedulers with same config names gives same results # for defaults lowerCAmelCase = UniPCMultistepScheduler(**self.get_scheduler_config() ) lowerCAmelCase = self.full_loop(scheduler=UpperCAmelCase_ ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 lowerCAmelCase = DPMSolverSinglestepScheduler.from_config(scheduler.config ) lowerCAmelCase = DEISMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = UniPCMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = self.full_loop(scheduler=UpperCAmelCase_ ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 def __snake_case ( self ): for timesteps in [25, 50, 1_00, 9_99, 10_00]: self.check_over_configs(num_train_timesteps=UpperCAmelCase_ ) def __snake_case ( self ): self.check_over_configs(thresholding=UpperCAmelCase_ ) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=UpperCAmelCase_ , prediction_type=UpperCAmelCase_ , sample_max_value=UpperCAmelCase_ , solver_order=UpperCAmelCase_ , solver_type=UpperCAmelCase_ , ) def __snake_case ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase_ ) def __snake_case ( self ): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=UpperCAmelCase_ , solver_type=UpperCAmelCase_ , prediction_type=UpperCAmelCase_ , ) lowerCAmelCase = self.full_loop( solver_order=UpperCAmelCase_ , solver_type=UpperCAmelCase_ , prediction_type=UpperCAmelCase_ , ) assert not torch.isnan(UpperCAmelCase_ ).any(), "Samples have nan numbers" def __snake_case ( self ): self.check_over_configs(lower_order_final=UpperCAmelCase_ ) self.check_over_configs(lower_order_final=UpperCAmelCase_ ) def __snake_case ( self ): for num_inference_steps in [1, 2, 3, 5, 10, 50, 1_00, 9_99, 10_00]: self.check_over_forward(num_inference_steps=UpperCAmelCase_ , time_step=0 ) def __snake_case ( self ): lowerCAmelCase = self.full_loop() lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 def __snake_case ( self ): lowerCAmelCase = self.full_loop(prediction_type='''v_prediction''' ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_mean.item() - 0.1014 ) < 1E-3 def __snake_case ( self ): lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(thresholding=UpperCAmelCase_ , dynamic_thresholding_ratio=0 ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) lowerCAmelCase = 10 lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter.half() scheduler.set_timesteps(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ).prev_sample assert sample.dtype == torch.floataa def __snake_case ( self , **UpperCAmelCase_ ): for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config(**UpperCAmelCase_ ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(scheduler.config.num_train_timesteps ) assert len(scheduler.timesteps.unique() ) == scheduler.num_inference_steps
702
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ =TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = self lowerCAmelCase = 0 class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # create a new set with x as its member lowerCAmelCase = DisjointSetTreeNode(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): # find the set x belongs to (with path-compression) lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # helper function for union operation if nodea.rank > nodea.rank: lowerCAmelCase = nodea else: lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # merge 2 disjoint sets self.link(self.find_set(UpperCAmelCase_ ) , self.find_set(UpperCAmelCase_ ) ) class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # add a node ONLY if its not present in the graph if node not in self.connections: lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # add an edge with the given weight self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) lowerCAmelCase = weight lowerCAmelCase = weight def __snake_case ( self ): lowerCAmelCase = [] lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCAmelCase_ : x[2] ) # creating the disjoint set lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(UpperCAmelCase_ ) # MST generation lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = edges[index] index += 1 lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) disjoint_set.union(UpperCAmelCase_ , UpperCAmelCase_ ) return graph
33
0
import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCAmelCase_ , '''width_multiplier''' ) ) class __UpperCamelCase : '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=64 , UpperCAmelCase_=2 , UpperCAmelCase_=3 , UpperCAmelCase_="swish" , UpperCAmelCase_=3 , UpperCAmelCase_=32 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.02 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=10 , UpperCAmelCase_=None , UpperCAmelCase_=0.25 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = make_divisible(5_12 * width_multiplier , divisor=8 ) lowerCAmelCase = hidden_act lowerCAmelCase = conv_kernel_size lowerCAmelCase = output_stride lowerCAmelCase = classifier_dropout_prob lowerCAmelCase = use_labels lowerCAmelCase = is_training lowerCAmelCase = num_labels lowerCAmelCase = initializer_range lowerCAmelCase = scope lowerCAmelCase = width_multiplier lowerCAmelCase = ffn_dropout lowerCAmelCase = attn_dropout def __snake_case ( self ): lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCAmelCase = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self ): return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = MobileViTVaModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = self.num_labels lowerCAmelCase = MobileViTVaForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = self.num_labels lowerCAmelCase = MobileViTVaForSemanticSegmentation(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCAmelCase = model(UpperCAmelCase_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowerCAmelCase = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self ): lowerCAmelCase = self.prepare_config_and_inputs() lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = config_and_inputs lowerCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : List[str] =( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __a : Union[str, Any] =( { """feature-extraction""": MobileViTVaModel, """image-classification""": MobileViTVaForImageClassification, """image-segmentation""": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __a : str =False __a : Dict =False __a : Union[str, Any] =False __a : Optional[int] =False def __snake_case ( self ): lowerCAmelCase = MobileViTVaModelTester(self ) lowerCAmelCase = MobileViTVaConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ ) def __snake_case ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''MobileViTV2 does not use inputs_embeds''' ) def __snake_case ( self ): pass @unittest.skip(reason='''MobileViTV2 does not support input and output embeddings''' ) def __snake_case ( self ): pass @unittest.skip(reason='''MobileViTV2 does not output attentions''' ) def __snake_case ( self ): pass @require_torch_multi_gpu @unittest.skip(reason='''Got `CUDA error: misaligned address` for tests after this one being run.''' ) def __snake_case ( self ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __snake_case ( self ): pass def __snake_case ( self ): lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(UpperCAmelCase_ ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def __snake_case ( self ): def check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() with torch.no_grad(): lowerCAmelCase = model(**self._prepare_for_class(UpperCAmelCase_ , UpperCAmelCase_ ) ) lowerCAmelCase = outputs.hidden_states lowerCAmelCase = 5 self.assertEqual(len(UpperCAmelCase_ ) , UpperCAmelCase_ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. lowerCAmelCase = 2 for i in range(len(UpperCAmelCase_ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = True check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase = True check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase_ ) @slow def __snake_case ( self ): for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = MobileViTVaModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCAmelCase ( ): lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __snake_case ( self ): return ( MobileViTImageProcessor.from_pretrained('''apple/mobilevitv2-1.0-imagenet1k-256''' ) if is_vision_available() else None ) @slow def __snake_case ( self ): lowerCAmelCase = MobileViTVaForImageClassification.from_pretrained('''apple/mobilevitv2-1.0-imagenet1k-256''' ).to( UpperCAmelCase_ ) lowerCAmelCase = self.default_image_processor lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**UpperCAmelCase_ ) # verify the logits lowerCAmelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([-1.6_3_3_6E0_0, -7.3_2_0_4E-0_2, -5.1_8_8_3E-0_1] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def __snake_case ( self ): lowerCAmelCase = MobileViTVaForSemanticSegmentation.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) lowerCAmelCase = model.to(UpperCAmelCase_ ) lowerCAmelCase = MobileViTImageProcessor.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**UpperCAmelCase_ ) lowerCAmelCase = outputs.logits # verify the logits lowerCAmelCase = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=UpperCAmelCase_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def __snake_case ( self ): lowerCAmelCase = MobileViTVaForSemanticSegmentation.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) lowerCAmelCase = model.to(UpperCAmelCase_ ) lowerCAmelCase = MobileViTImageProcessor.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) lowerCAmelCase = prepare_img() lowerCAmelCase = image_processor(images=UpperCAmelCase_ , return_tensors='''pt''' ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase = model(**UpperCAmelCase_ ) lowerCAmelCase = outputs.logits.detach().cpu() lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=UpperCAmelCase_ , target_sizes=[(50, 60)] ) lowerCAmelCase = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , UpperCAmelCase_ ) lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=UpperCAmelCase_ ) lowerCAmelCase = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , UpperCAmelCase_ )
703
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations(_snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations_with_dp_array( _snake_case , _snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _snake_case ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =3 UpperCAmelCase_ =5 UpperCAmelCase_ =[1, 2, 5] print(combination_sum_iv(n, array, target))
33
0
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __UpperCamelCase : '''simple docstring''' __a : Dict =BlenderbotConfig __a : Dict ={} __a : Dict ="""gelu""" def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=7 , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=99 , UpperCAmelCase_=32 , UpperCAmelCase_=2 , UpperCAmelCase_=4 , UpperCAmelCase_=37 , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=2 , UpperCAmelCase_=1 , UpperCAmelCase_=0 , ): lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = eos_token_id lowerCAmelCase = pad_token_id lowerCAmelCase = bos_token_id def __snake_case ( self ): lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowerCAmelCase = prepare_blenderbot_inputs_dict(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return config, inputs_dict def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = TFBlenderbotModel(config=UpperCAmelCase_ ).get_decoder() lowerCAmelCase = inputs_dict['''input_ids'''] lowerCAmelCase = input_ids[:1, :] lowerCAmelCase = inputs_dict['''attention_mask'''][:1, :] lowerCAmelCase = inputs_dict['''head_mask'''] lowerCAmelCase = 1 # first forward pass lowerCAmelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , head_mask=UpperCAmelCase_ , use_cache=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCAmelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] lowerCAmelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , past_key_values=UpperCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(UpperCAmelCase_ , UpperCAmelCase_ , rtol=1E-3 ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case=None , ): if attention_mask is None: lowerCAmelCase = tf.cast(tf.math.not_equal(_snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =(TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __a : Union[str, Any] =(TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __a : int =( { """conversational""": TFBlenderbotForConditionalGeneration, """feature-extraction""": TFBlenderbotModel, """summarization""": TFBlenderbotForConditionalGeneration, """text2text-generation""": TFBlenderbotForConditionalGeneration, """translation""": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __a : List[Any] =True __a : Optional[int] =False __a : Union[str, Any] =False def __snake_case ( self ): lowerCAmelCase = TFBlenderbotModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=UpperCAmelCase_ ) def __snake_case ( self ): self.config_tester.run_common_tests() def __snake_case ( self ): lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*UpperCAmelCase_ ) @require_tokenizers @require_tf class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : Dict =["""My friends are cool but they eat too many carbs."""] __a : Union[str, Any] ="""facebook/blenderbot-400M-distill""" @cached_property def __snake_case ( self ): return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def __snake_case ( self ): lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def __snake_case ( self ): lowerCAmelCase = self.tokenizer(self.src_text , return_tensors='''tf''' ) lowerCAmelCase = self.model.generate( model_inputs.input_ids , ) lowerCAmelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=UpperCAmelCase_ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
704
import torch from diffusers import StableDiffusionPipeline UpperCAmelCase_ ="""path-to-your-trained-model""" UpperCAmelCase_ =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") UpperCAmelCase_ ="""A photo of sks dog in a bucket""" UpperCAmelCase_ =pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
33
0
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : torch.FloatTensor __a : torch.FloatTensor class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : int =1 @register_to_config def __init__( self , UpperCAmelCase_ = 20_00 , UpperCAmelCase_ = 0.15 , UpperCAmelCase_ = 0.01 , UpperCAmelCase_ = 1348.0 , UpperCAmelCase_ = 1E-5 , UpperCAmelCase_ = 1 , ): # standard deviation of the initial noise distribution lowerCAmelCase = sigma_max # setable values lowerCAmelCase = None self.set_sigmas(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): return sample def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None ): lowerCAmelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps lowerCAmelCase = torch.linspace(1 , UpperCAmelCase_ , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = None ): lowerCAmelCase = sigma_min if sigma_min is not None else self.config.sigma_min lowerCAmelCase = sigma_max if sigma_max is not None else self.config.sigma_max lowerCAmelCase = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) lowerCAmelCase = torch.exp(torch.linspace(math.log(UpperCAmelCase_ ) , math.log(UpperCAmelCase_ ) , UpperCAmelCase_ ) ) lowerCAmelCase = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = True , ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) lowerCAmelCase = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) lowerCAmelCase = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda lowerCAmelCase = timesteps.to(self.discrete_sigmas.device ) lowerCAmelCase = self.discrete_sigmas[timesteps].to(sample.device ) lowerCAmelCase = self.get_adjacent_sigma(UpperCAmelCase_ , UpperCAmelCase_ ).to(sample.device ) lowerCAmelCase = torch.zeros_like(UpperCAmelCase_ ) lowerCAmelCase = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods lowerCAmelCase = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): lowerCAmelCase = diffusion.unsqueeze(-1 ) lowerCAmelCase = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of lowerCAmelCase = randn_tensor( sample.shape , layout=sample.layout , generator=UpperCAmelCase_ , device=sample.device , dtype=sample.dtype ) lowerCAmelCase = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? lowerCAmelCase = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=UpperCAmelCase_ , prev_sample_mean=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = True , ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction lowerCAmelCase = randn_tensor(sample.shape , layout=sample.layout , generator=UpperCAmelCase_ ).to(sample.device ) # compute step size from the model_output, the noise, and the snr lowerCAmelCase = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() lowerCAmelCase = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() lowerCAmelCase = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 lowerCAmelCase = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term lowerCAmelCase = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): lowerCAmelCase = step_size.unsqueeze(-1 ) lowerCAmelCase = sample + step_size * model_output lowerCAmelCase = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ): # Make sure sigmas and timesteps have the same device and dtype as original_samples lowerCAmelCase = timesteps.to(original_samples.device ) lowerCAmelCase = self.discrete_sigmas.to(original_samples.device )[timesteps] lowerCAmelCase = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(UpperCAmelCase_ ) * sigmas[:, None, None, None] ) lowerCAmelCase = noise + original_samples return noisy_samples def __len__( self ): return self.config.num_train_timesteps
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
from __future__ import annotations import math from collections.abc import Callable def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case = 100 , ): lowerCAmelCase = x_start lowerCAmelCase = fnc(_snake_case ) lowerCAmelCase = 0.0 for _ in range(_snake_case ): # Approximates curve as a sequence of linear lines and sums their length lowerCAmelCase = (x_end - x_start) / steps + xa lowerCAmelCase = fnc(_snake_case ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowerCAmelCase = xa lowerCAmelCase = fxa return length if __name__ == "__main__": def UpperCAmelCase ( _snake_case ): return math.sin(10 * x ) print("""f(x) = sin(10 * x)""") print("""The length of the curve from x = -10 to x = 10 is:""") UpperCAmelCase_ =10 while i <= 10_0000: print(F'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
706
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = 8 # DPR tok lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __snake_case ( self ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __snake_case ( self ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: lowerCAmelCase = os.path.join(self.tmpdirname , '''dataset''' ) lowerCAmelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCAmelCase_ ) , ) return retriever def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) lowerCAmelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) lowerCAmelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(UpperCAmelCase_ , open(UpperCAmelCase_ , '''wb''' ) ) lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = self.get_dummy_dataset() retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_legacy_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): import torch lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) lowerCAmelCase = retriever( UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ , return_tensors='''pt''' , ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): lowerCAmelCase = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) retriever.set_ctx_encoder_tokenizer(UpperCAmelCase_ ) lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) self.assertEqual( len(UpperCAmelCase_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCAmelCase_ ) # check for doc token related keys in dictionary.
33
0
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ =2048 UpperCAmelCase_ =4096 UpperCAmelCase_ =42 UpperCAmelCase_ =os.environ.pop("""PROCESS_TRAIN""", """false""") UpperCAmelCase_ ={"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def UpperCAmelCase ( _snake_case ): def choose_first(_snake_case , _snake_case=False ): assert isinstance(_snake_case , _snake_case ) if len(_snake_case ) == 1: lowerCAmelCase = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: lowerCAmelCase = {k: [a[k]] for k in a} if len(a['''start_token'''] ) > 0: break return a lowerCAmelCase = {'''id''': example['''id''']} lowerCAmelCase = example['''annotations'''] lowerCAmelCase = annotation['''yes_no_answer'''] if 0 in yes_no_answer or 1 in yes_no_answer: lowerCAmelCase = ['''yes'''] if 1 in yes_no_answer else ['''no'''] lowerCAmelCase = lowerCAmelCase = [] lowerCAmelCase = lowerCAmelCase = [] lowerCAmelCase = ['''<cls>'''] else: lowerCAmelCase = ['''short'''] lowerCAmelCase = choose_first(annotation['''short_answers'''] ) if len(out['''start_token'''] ) == 0: # answer will be long if short is not available lowerCAmelCase = ['''long'''] lowerCAmelCase = choose_first(annotation['''long_answer'''] , is_long_answer=_snake_case ) lowerCAmelCase = [] answer.update(_snake_case ) # disregard some samples if len(answer['''start_token'''] ) > 1 or answer["start_token"] == answer["end_token"]: lowerCAmelCase = True else: lowerCAmelCase = False lowerCAmelCase = ['''start_token''', '''end_token''', '''start_byte''', '''end_byte''', '''text'''] if not all(isinstance(answer[k] , _snake_case ) for k in cols ): raise ValueError('''Issue in ID''' , example['''id'''] ) return answer def UpperCAmelCase ( _snake_case , _snake_case=False ): lowerCAmelCase = _get_single_answer(_snake_case ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase = example['''document''']['''tokens'''] lowerCAmelCase = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) return { "context": " ".join(_snake_case ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples lowerCAmelCase = ['''start_token''', '''end_token'''] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 lowerCAmelCase = example['''document''']['''tokens'''] lowerCAmelCase = answer['''start_token'''] lowerCAmelCase = answer['''end_token'''] lowerCAmelCase = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 lowerCAmelCase = ''' '''.join(context[start_token:end_token] ) # checking above code if assertion: lowerCAmelCase = doc['''is_html'''][answer['''start_token'''] : answer['''end_token''']] lowerCAmelCase = doc['''token'''][answer['''start_token'''] : answer['''end_token''']] lowerCAmelCase = ''' '''.join([old[i] for i in range(len(_snake_case ) ) if not is_html[i]] ) if new != old: print('''ID:''' , example['''id'''] ) print('''New:''' , _snake_case , end='''\n''' ) print('''Old:''' , _snake_case , end='''\n\n''' ) return { "context": " ".join(_snake_case ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def UpperCAmelCase ( _snake_case , _snake_case , _snake_case=2048 , _snake_case=4096 , _snake_case=True ): # overlap will be of doc_stride - q_len lowerCAmelCase = get_context_and_ans(_snake_case , assertion=_snake_case ) lowerCAmelCase = out['''answer'''] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } lowerCAmelCase = tokenizer(example['''question''']['''text'''] , out['''context'''] ).input_ids lowerCAmelCase = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = input_ids[:q_len] lowerCAmelCase = range(_snake_case , len(_snake_case ) , max_length - doc_stride ) for i in doc_start_indices: lowerCAmelCase = i + max_length - q_len lowerCAmelCase = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['''category'''][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(_snake_case ), "end_token": [-100] * len(_snake_case ), "category": category, }, } lowerCAmelCase = out['''context'''].split() lowerCAmelCase = splitted_context[answer['''end_token''']] lowerCAmelCase = len( tokenizer( ''' '''.join(splitted_context[: answer['''start_token''']] ) , add_special_tokens=_snake_case , ).input_ids ) lowerCAmelCase = len( tokenizer(''' '''.join(splitted_context[: answer['''end_token''']] ) , add_special_tokens=_snake_case ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token lowerCAmelCase = len(tokenizer(_snake_case , add_special_tokens=_snake_case ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 lowerCAmelCase = input_ids[answer['''start_token'''] : answer['''end_token'''] + 1] # right & left are inclusive lowerCAmelCase = answer['''start_token'''] lowerCAmelCase = answer['''end_token'''] if assertion: lowerCAmelCase = tokenizer.decode(_snake_case ) if answer["span"] != new: print('''ISSUE IN TOKENIZATION''' ) print('''OLD:''' , answer['''span'''] ) print('''NEW:''' , _snake_case , end='''\n\n''' ) if len(_snake_case ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } lowerCAmelCase = input_ids[:q_len] lowerCAmelCase = range(_snake_case , len(_snake_case ) , max_length - doc_stride ) lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = [] # null, yes, no, long, short for i in doc_start_indices: lowerCAmelCase = i + max_length - q_len lowerCAmelCase = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: lowerCAmelCase = start_token - i + q_len lowerCAmelCase = end_token - i + q_len answers_category.append(answer['''category'''][0] ) # ["short"] -> "short" else: lowerCAmelCase = -100 lowerCAmelCase = -100 answers_category.append('''null''' ) lowerCAmelCase = inputs[-1][start_token : end_token + 1] answers_start_token.append(_snake_case ) answers_end_token.append(_snake_case ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('''ISSUE in strided for ID:''' , example['''id'''] ) print('''New:''' , tokenizer.decode(_snake_case ) ) print('''Old:''' , tokenizer.decode(_snake_case ) , end='''\n\n''' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def UpperCAmelCase ( _snake_case , _snake_case , _snake_case=2048 , _snake_case=4096 , _snake_case=False ): lowerCAmelCase = get_strided_contexts_and_ans( _snake_case , _snake_case , doc_stride=_snake_case , max_length=_snake_case , assertion=_snake_case , ) return example def UpperCAmelCase ( _snake_case , _snake_case ): with jsonlines.open(_snake_case , '''a''' ) as writer: for example in tqdm(_snake_case , total=len(_snake_case ) , desc='''Saving samples ... ''' ): lowerCAmelCase = example['''labels'''] for ids, start, end, cat in zip( example['''input_ids'''] , labels['''start_token'''] , labels['''end_token'''] , labels['''category'''] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { '''input_ids''': ids, '''start_token''': start, '''end_token''': end, '''category''': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer UpperCAmelCase_ =load_dataset("""natural_questions""") UpperCAmelCase_ =BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") UpperCAmelCase_ =data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] UpperCAmelCase_ ={ """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } UpperCAmelCase_ =data.map(prepare_inputs, fn_kwargs=fn_kwargs) UpperCAmelCase_ =data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) UpperCAmelCase_ ="""nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
707
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """google/switch-base-8""": """https://huggingface.co/google/switch-base-8/blob/main/config.json""", } class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ="""switch_transformers""" __a : Union[str, Any] =["""past_key_values"""] __a : Dict ={"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , UpperCAmelCase_=3_21_28 , UpperCAmelCase_=7_68 , UpperCAmelCase_=64 , UpperCAmelCase_=20_48 , UpperCAmelCase_=64 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=3 , UpperCAmelCase_=12 , UpperCAmelCase_=8 , UpperCAmelCase_=False , UpperCAmelCase_=0.01 , UpperCAmelCase_="float32" , UpperCAmelCase_=False , UpperCAmelCase_=32 , UpperCAmelCase_=1_28 , UpperCAmelCase_=0.1 , UpperCAmelCase_=1E-6 , UpperCAmelCase_=0.001 , UpperCAmelCase_=0.001 , UpperCAmelCase_=1.0 , UpperCAmelCase_="relu" , UpperCAmelCase_=True , UpperCAmelCase_=False , UpperCAmelCase_=True , UpperCAmelCase_=0 , UpperCAmelCase_=1 , **UpperCAmelCase_ , ): lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = d_kv lowerCAmelCase = d_ff lowerCAmelCase = num_sparse_encoder_layers lowerCAmelCase = num_layers lowerCAmelCase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: lowerCAmelCase = self.num_layers // self.num_sparse_encoder_layers else: lowerCAmelCase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: lowerCAmelCase = self.num_decoder_layers // self.num_sparse_decoder_layers else: lowerCAmelCase = self.num_decoder_layers # HACK: this will create 0 sparse layers lowerCAmelCase = num_heads lowerCAmelCase = num_experts lowerCAmelCase = expert_capacity lowerCAmelCase = router_bias lowerCAmelCase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) lowerCAmelCase = router_dtype lowerCAmelCase = router_ignore_padding_tokens lowerCAmelCase = relative_attention_num_buckets lowerCAmelCase = relative_attention_max_distance lowerCAmelCase = dropout_rate lowerCAmelCase = layer_norm_epsilon lowerCAmelCase = initializer_factor lowerCAmelCase = feed_forward_proj lowerCAmelCase = use_cache lowerCAmelCase = add_router_probs lowerCAmelCase = router_z_loss_coef lowerCAmelCase = router_aux_loss_coef lowerCAmelCase = self.feed_forward_proj.split('''-''' ) lowerCAmelCase = act_info[-1] lowerCAmelCase = act_info[0] == '''gated''' if len(UpperCAmelCase_ ) > 1 and act_info[0] != "gated" or len(UpperCAmelCase_ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCAmelCase = '''gelu_new''' super().__init__( pad_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , is_encoder_decoder=UpperCAmelCase_ , **UpperCAmelCase_ , )
33
0
from __future__ import annotations def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [] lowerCAmelCase , lowerCAmelCase = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) lowerCAmelCase = result + left + right return input_list def UpperCAmelCase ( _snake_case ): if len(_snake_case ) <= 1: return input_list lowerCAmelCase = list(_snake_case ) # iteration for two-way merging lowerCAmelCase = 2 while p <= len(_snake_case ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(_snake_case ) , _snake_case ): lowerCAmelCase = i lowerCAmelCase = i + p - 1 lowerCAmelCase = (low + high + 1) // 2 lowerCAmelCase = merge(_snake_case , _snake_case , _snake_case , _snake_case ) # final merge of last two parts if p * 2 >= len(_snake_case ): lowerCAmelCase = i lowerCAmelCase = merge(_snake_case , 0 , _snake_case , len(_snake_case ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": UpperCAmelCase_ =input("""Enter numbers separated by a comma:\n""").strip() if user_input == "": UpperCAmelCase_ =[] else: UpperCAmelCase_ =[int(item.strip()) for item in user_input.split(""",""")] print(iter_merge_sort(unsorted))
708
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def UpperCAmelCase ( _snake_case ): lowerCAmelCase , lowerCAmelCase = analyze_text(_snake_case ) lowerCAmelCase = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase = sum(single_char_strings.values() ) # one length string lowerCAmelCase = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase = single_char_strings[ch] lowerCAmelCase = my_str / all_sum my_fir_sum += prob * math.loga(_snake_case ) # entropy formula. # print entropy print(F"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase = sum(two_char_strings.values() ) lowerCAmelCase = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase = cha + cha if sequence in two_char_strings: lowerCAmelCase = two_char_strings[sequence] lowerCAmelCase = int(_snake_case ) / all_sum my_sec_sum += prob * math.loga(_snake_case ) # print second entropy print(F"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(F"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def UpperCAmelCase ( _snake_case ): lowerCAmelCase = Counter() # type: ignore lowerCAmelCase = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(_snake_case ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def UpperCAmelCase ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
33
0
import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = 0 @slow def __snake_case ( self ): for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(UpperCAmelCase_ ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(UpperCAmelCase_ ) , 0 ) def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 20 ) def __snake_case ( self ): lowerCAmelCase = AutoConfig.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) # Check that tokenizer_type ≠ model_type lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , config=UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('''./tests/fixtures/vocab.txt''' , os.path.join(UpperCAmelCase_ , '''vocab.txt''' ) ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , tokenizer_type='''bert''' , use_fast=UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('''./tests/fixtures/vocab.json''' , os.path.join(UpperCAmelCase_ , '''vocab.json''' ) ) shutil.copy('''./tests/fixtures/merges.txt''' , os.path.join(UpperCAmelCase_ , '''merges.txt''' ) ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , tokenizer_type='''gpt2''' , use_fast=UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) @require_tokenizers def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('''./tests/fixtures/vocab.txt''' , os.path.join(UpperCAmelCase_ , '''vocab.txt''' ) ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , tokenizer_type='''bert''' ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('''./tests/fixtures/vocab.json''' , os.path.join(UpperCAmelCase_ , '''vocab.json''' ) ) shutil.copy('''./tests/fixtures/merges.txt''' , os.path.join(UpperCAmelCase_ , '''merges.txt''' ) ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , tokenizer_type='''gpt2''' ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): with pytest.raises(UpperCAmelCase_ ): AutoTokenizer.from_pretrained('''./''' , tokenizer_type='''xxx''' ) @require_tokenizers def __snake_case ( self ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: lowerCAmelCase = tokenizer_class.from_pretrained('''wietsedv/bert-base-dutch-cased''' ) self.assertIsInstance(UpperCAmelCase_ , (BertTokenizer, BertTokenizerFast) ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , UpperCAmelCase_ ) else: self.assertEqual(tokenizer.do_lower_case , UpperCAmelCase_ ) self.assertEqual(tokenizer.model_max_length , 5_12 ) @require_tokenizers def __snake_case ( self ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( UpperCAmelCase_ , '''julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier''' , ): lowerCAmelCase = tokenizer_class.from_pretrained('''julien-c/herlolip-not-exists''' ) def __snake_case ( self ): # tests: https://github.com/huggingface/transformers/pull/13251 # 1. models with `-`, e.g. xlm-roberta -> xlm_roberta # 2. models that don't remap 1-1 from model-name to model file, e.g., openai-gpt -> openai lowerCAmelCase = TOKENIZER_MAPPING.values() lowerCAmelCase = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(UpperCAmelCase_ ) @require_tokenizers def __snake_case ( self ): self.assertIsInstance(AutoTokenizer.from_pretrained('''bert-base-cased''' , use_fast=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertIsInstance(AutoTokenizer.from_pretrained('''bert-base-cased''' ) , UpperCAmelCase_ ) @require_tokenizers def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained('''distilbert-base-uncased''' , do_lower_case=UpperCAmelCase_ ) lowerCAmelCase = '''Hello, world. How are you?''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertEqual('''[UNK]''' , tokens[0] ) lowerCAmelCase = AutoTokenizer.from_pretrained('''microsoft/mpnet-base''' , do_lower_case=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertEqual('''[UNK]''' , tokens[0] ) @require_tokenizers def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained('''robot-test/dummy-tokenizer-fast-with-model-config''' ) self.assertEqual(type(UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(tokenizer.model_max_length , 5_12 ) self.assertEqual(tokenizer.vocab_size , 3_00_00 ) self.assertEqual(tokenizer.unk_token , '''[UNK]''' ) self.assertEqual(tokenizer.padding_side , '''right''' ) self.assertEqual(tokenizer.truncation_side , '''right''' ) def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 12 ) def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained('''ctrl''' ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): # Check we can load the tokenizer config of an online model. lowerCAmelCase = get_tokenizer_config('''bert-base-cased''' ) lowerCAmelCase = config.pop('''_commit_hash''' , UpperCAmelCase_ ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(UpperCAmelCase_ , {'''do_lower_case''': False} ) # This model does not have a tokenizer_config so we get back an empty dict. lowerCAmelCase = get_tokenizer_config(UpperCAmelCase_ ) self.assertDictEqual(UpperCAmelCase_ , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = get_tokenizer_config(UpperCAmelCase_ ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config['''tokenizer_class'''] , '''BertTokenizer''' ) def __snake_case ( self ): try: AutoConfig.register('''custom''' , UpperCAmelCase_ ) AutoTokenizer.register(UpperCAmelCase_ , slow_tokenizer_class=UpperCAmelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCAmelCase_ ): AutoTokenizer.register(UpperCAmelCase_ , slow_tokenizer_class=UpperCAmelCase_ ) lowerCAmelCase = CustomTokenizer.from_pretrained(UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def __snake_case ( self ): try: AutoConfig.register('''custom''' , UpperCAmelCase_ ) # Can register in two steps AutoTokenizer.register(UpperCAmelCase_ , slow_tokenizer_class=UpperCAmelCase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(UpperCAmelCase_ , fast_tokenizer_class=UpperCAmelCase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( UpperCAmelCase_ , slow_tokenizer_class=UpperCAmelCase_ , fast_tokenizer_class=UpperCAmelCase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCAmelCase_ ): AutoTokenizer.register(UpperCAmelCase_ , fast_tokenizer_class=UpperCAmelCase_ ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase = BertTokenizerFast.from_pretrained(UpperCAmelCase_ ) bert_tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = CustomTokenizerFast.from_pretrained(UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def __snake_case ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(UpperCAmelCase_ ): lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/test_dynamic_tokenizer''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(UpperCAmelCase_ ): lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , trust_remote_code=UpperCAmelCase_ ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) # Test we can also load the slow version lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , trust_remote_code=UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , '''NewTokenizer''' ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , '''NewTokenizer''' ) @require_tokenizers def __snake_case ( self ): class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any =False class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : List[str] =NewTokenizer __a : List[Any] =False try: AutoConfig.register('''custom''' , UpperCAmelCase_ ) AutoTokenizer.register(UpperCAmelCase_ , slow_tokenizer_class=UpperCAmelCase_ ) AutoTokenizer.register(UpperCAmelCase_ , fast_tokenizer_class=UpperCAmelCase_ ) # If remote code is not set, the default is to use local lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/test_dynamic_tokenizer''' ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) self.assertFalse(tokenizer.special_attribute_present ) lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/test_dynamic_tokenizer''' , use_fast=UpperCAmelCase_ ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) self.assertFalse(tokenizer.special_attribute_present ) lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) self.assertTrue(tokenizer.special_attribute_present ) lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer''' , trust_remote_code=UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def __snake_case ( self ): lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer_legacy''' , trust_remote_code=UpperCAmelCase_ ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) # Test we can also load the slow version lowerCAmelCase = AutoTokenizer.from_pretrained( '''hf-internal-testing/test_dynamic_tokenizer_legacy''' , trust_remote_code=UpperCAmelCase_ , use_fast=UpperCAmelCase_ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) else: self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) def __snake_case ( self ): with self.assertRaisesRegex( UpperCAmelCase_ , '''bert-base is not a local folder and is not a valid model identifier''' ): lowerCAmelCase = AutoTokenizer.from_pretrained('''bert-base''' ) def __snake_case ( self ): with self.assertRaisesRegex( UpperCAmelCase_ , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ , revision='''aaaaaa''' ) def __snake_case ( self ): # Make sure we have cached the tokenizer. lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) with RequestCounter() as counter: lowerCAmelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
709
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Tuple =IFInpaintingSuperResolutionPipeline __a : Dict =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __a : int =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) __a : Union[str, Any] =PipelineTesterMixin.required_optional_params - {"""latents"""} def __snake_case ( self ): return self._get_superresolution_dummy_components() def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=0 ): if str(UpperCAmelCase_ ).startswith('''mps''' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase_ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __snake_case ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __snake_case ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __snake_case ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __snake_case ( self ): self._test_save_load_local() def __snake_case ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
33
0
import argparse import os import re import packaging.version UpperCAmelCase_ ="""examples/""" UpperCAmelCase_ ={ """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } UpperCAmelCase_ ={ """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } UpperCAmelCase_ ="""README.md""" def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): with open(_snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase = f.read() lowerCAmelCase , lowerCAmelCase = REPLACE_PATTERNS[pattern] lowerCAmelCase = replace.replace('''VERSION''' , _snake_case ) lowerCAmelCase = re_pattern.sub(_snake_case , _snake_case ) with open(_snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_snake_case ) def UpperCAmelCase ( _snake_case ): for folder, directories, fnames in os.walk(_snake_case ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_snake_case , _snake_case ) , _snake_case , pattern='''examples''' ) def UpperCAmelCase ( _snake_case , _snake_case=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_snake_case , _snake_case , _snake_case ) if not patch: update_version_in_examples(_snake_case ) def UpperCAmelCase ( ): lowerCAmelCase = '''🤗 Transformers currently provides the following architectures''' lowerCAmelCase = '''1. Want to contribute a new model?''' with open(_snake_case , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCAmelCase = f.readlines() # Find the start of the list. lowerCAmelCase = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCAmelCase = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowerCAmelCase = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(_snake_case , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_snake_case ) def UpperCAmelCase ( ): with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowerCAmelCase = f.read() lowerCAmelCase = REPLACE_PATTERNS['''init'''][0].search(_snake_case ).groups()[0] return packaging.version.parse(_snake_case ) def UpperCAmelCase ( _snake_case=False ): lowerCAmelCase = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowerCAmelCase = default_version.base_version elif patch: lowerCAmelCase = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowerCAmelCase = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowerCAmelCase = input(F"""Which version are you releasing? [{default_version}]""" ) if len(_snake_case ) == 0: lowerCAmelCase = default_version print(F"""Updating version to {version}.""" ) global_version_update(_snake_case , patch=_snake_case ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def UpperCAmelCase ( ): lowerCAmelCase = get_version() lowerCAmelCase = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowerCAmelCase = current_version.base_version # Check with the user we got that right. lowerCAmelCase = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(_snake_case ) == 0: lowerCAmelCase = dev_version print(F"""Updating version to {version}.""" ) global_version_update(_snake_case ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") UpperCAmelCase_ =parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
710
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCAmelCase_ ={ """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
'''simple docstring''' import numpy as np def UpperCAmelCase ( _snake_case , _snake_case ): return np.where(vector > 0 , _snake_case , (alpha * (np.exp(_snake_case ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
711
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline UpperCAmelCase_ =datasets.utils.logging.get_logger(__name__) @dataclass class __UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __a : Optional[datasets.Features] =None __a : str ="utf-8" __a : Optional[str] =None __a : Optional[str] =None __a : bool =True # deprecated __a : Optional[int] =None # deprecated __a : int =1_0 << 2_0 # 10MB __a : Optional[bool] =None class __UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __a : str =JsonConfig def __snake_case ( self ): if self.config.block_size is not None: logger.warning('''The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead''' ) lowerCAmelCase = self.config.block_size if self.config.use_threads is not True: logger.warning( '''The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.''' ) if self.config.newlines_in_values is not None: raise ValueError('''The JSON loader parameter `newlines_in_values` is no longer supported''' ) return datasets.DatasetInfo(features=self.config.features ) def __snake_case ( self , UpperCAmelCase_ ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase_ , (str, list, tuple) ): lowerCAmelCase = data_files if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = [files] lowerCAmelCase = [dl_manager.iter_files(UpperCAmelCase_ ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase_ , gen_kwargs={'''files''': files} ) ) return splits def __snake_case ( self , UpperCAmelCase_ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCAmelCase = self.config.features.arrow_schema.field(UpperCAmelCase_ ).type lowerCAmelCase = pa_table.append_column(UpperCAmelCase_ , pa.array([None] * len(UpperCAmelCase_ ) , type=UpperCAmelCase_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCAmelCase = table_cast(UpperCAmelCase_ , self.config.features.arrow_schema ) return pa_table def __snake_case ( self , UpperCAmelCase_ ): for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) # We keep only the field we are interested in lowerCAmelCase = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(UpperCAmelCase_ , (list, tuple) ): lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} else: lowerCAmelCase = dataset lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) yield file_idx, self._cast_table(UpperCAmelCase_ ) # If the file has one json object per line else: with open(UpperCAmelCase_ , '''rb''' ) as f: lowerCAmelCase = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCAmelCase = max(self.config.chunksize // 32 , 16 << 10 ) lowerCAmelCase = ( self.config.encoding_errors if self.config.encoding_errors is not None else '''strict''' ) while True: lowerCAmelCase = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(UpperCAmelCase_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCAmelCase = batch.decode(self.config.encoding , errors=UpperCAmelCase_ ).encode('''utf-8''' ) try: while True: try: lowerCAmelCase = paj.read_json( io.BytesIO(UpperCAmelCase_ ) , read_options=paj.ReadOptions(block_size=UpperCAmelCase_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(UpperCAmelCase_ , pa.ArrowInvalid ) and "straddling" not in str(UpperCAmelCase_ ) or block_size > len(UpperCAmelCase_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(UpperCAmelCase_ )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( UpperCAmelCase_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCAmelCase = json.load(UpperCAmelCase_ ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # list is the only sequence type supported in JSON try: lowerCAmelCase = set().union(*[row.keys() for row in dataset] ) lowerCAmelCase = {col: [row.get(UpperCAmelCase_ ) for row in dataset] for col in keys} lowerCAmelCase = pa.Table.from_pydict(UpperCAmelCase_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(UpperCAmelCase_ ) break else: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase_ )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase_ ) batch_idx += 1
33
0
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
712
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase_ =logging.get_logger(__name__) class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Optional[Any] ="""maskformer-swin""" __a : Optional[int] ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , UpperCAmelCase_=2_24 , UpperCAmelCase_=4 , UpperCAmelCase_=3 , UpperCAmelCase_=96 , UpperCAmelCase_=[2, 2, 6, 2] , UpperCAmelCase_=[3, 6, 12, 24] , UpperCAmelCase_=7 , UpperCAmelCase_=4.0 , UpperCAmelCase_=True , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.0 , UpperCAmelCase_=0.1 , UpperCAmelCase_="gelu" , UpperCAmelCase_=False , UpperCAmelCase_=0.02 , UpperCAmelCase_=1E-5 , UpperCAmelCase_=None , UpperCAmelCase_=None , **UpperCAmelCase_ , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase = image_size lowerCAmelCase = patch_size lowerCAmelCase = num_channels lowerCAmelCase = embed_dim lowerCAmelCase = depths lowerCAmelCase = len(UpperCAmelCase_ ) lowerCAmelCase = num_heads lowerCAmelCase = window_size lowerCAmelCase = mlp_ratio lowerCAmelCase = qkv_bias lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = drop_path_rate lowerCAmelCase = hidden_act lowerCAmelCase = use_absolute_embeddings lowerCAmelCase = layer_norm_eps lowerCAmelCase = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase = int(embed_dim * 2 ** (len(UpperCAmelCase_ ) - 1) ) lowerCAmelCase = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase_ ) + 1 )] lowerCAmelCase , lowerCAmelCase = get_aligned_output_features_output_indices( out_features=UpperCAmelCase_ , out_indices=UpperCAmelCase_ , stage_names=self.stage_names )
33
0
from cva import destroyAllWindows, imread, imshow, waitKey def UpperCAmelCase ( _snake_case ): # getting number of pixels in the image lowerCAmelCase , lowerCAmelCase = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(_snake_case ): for j in range(_snake_case ): lowerCAmelCase = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image UpperCAmelCase_ =imread("""image_data/lena.jpg""", 1) # convert to its negative UpperCAmelCase_ =convert_to_negative(img) # show result image imshow("""negative of original image""", img) waitKey(0) destroyAllWindows()
713
from collections.abc import Sequence def UpperCAmelCase ( _snake_case , _snake_case = False ): if not arr: return 0 lowerCAmelCase = 0 if allow_empty_subarrays else float('''-inf''' ) lowerCAmelCase = 0.0 for num in arr: lowerCAmelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) lowerCAmelCase = max(_snake_case , _snake_case ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase_ =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'''{max_subarray_sum(nums) = }''')
33
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase_ ={"""configuration_unispeech""": ["""UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP""", """UniSpeechConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST""", """UniSpeechForCTC""", """UniSpeechForPreTraining""", """UniSpeechForSequenceClassification""", """UniSpeechModel""", """UniSpeechPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
714
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any =BertJapaneseTokenizer __a : Optional[int] =False __a : int =True def __snake_case ( self ): super().setUp() lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase = self.get_input_output_texts(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(do_lower_case=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __snake_case ( self ): try: lowerCAmelCase = MecabTokenizer( do_lower_case=UpperCAmelCase_ , normalize_text=UpperCAmelCase_ , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = MecabTokenizer(normalize_text=UpperCAmelCase_ , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(do_lower_case=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(normalize_text=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __snake_case ( self ): lowerCAmelCase = SudachiTokenizer(trim_whitespace=UpperCAmelCase_ , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(UpperCAmelCase_ ) lowerCAmelCase = '''こんにちは、世界。\nこんばんは、世界。''' lowerCAmelCase = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) lowerCAmelCase = os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(UpperCAmelCase_ , '''wb''' ) as handle: pickle.dump(UpperCAmelCase_ , UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''rb''' ) as handle: lowerCAmelCase = pickle.load(UpperCAmelCase_ ) lowerCAmelCase = tokenizer_new.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(do_lower_case=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(normalize_text=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer(trim_whitespace=UpperCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __snake_case ( self ): lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = WordpieceTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) lowerCAmelCase = tokenizer.subword_tokenizer lowerCAmelCase = subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) lowerCAmelCase = subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(UpperCAmelCase_ , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =BertJapaneseTokenizer __a : Optional[int] =False def __snake_case ( self ): super().setUp() lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __snake_case ( self , **UpperCAmelCase_ ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = '''こんにちは、世界。 \nこんばんは、世界。''' lowerCAmelCase = '''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): pass # TODO add if relevant def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) lowerCAmelCase = tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( UpperCAmelCase_ , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def __snake_case ( self ): lowerCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] lowerCAmelCase = {} for i, token in enumerate(UpperCAmelCase_ ): lowerCAmelCase = i lowerCAmelCase = CharacterTokenizer(vocab=UpperCAmelCase_ , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __snake_case ( self ): lowerCAmelCase = self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) lowerCAmelCase = tokenizer.encode('''ありがとう。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.encode('''どういたしまして。''' , add_special_tokens=UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ , UpperCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' lowerCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = '''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) lowerCAmelCase = '''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(UpperCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
33
0
def UpperCAmelCase ( _snake_case = 100 ): '''simple docstring''' lowerCAmelCase = (n * (n + 1) // 2) ** 2 lowerCAmelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(F'''{solution() = }''')
715
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) UpperCAmelCase_ ="""hf-internal-testing/tiny-random-bert""" UpperCAmelCase_ =os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") UpperCAmelCase_ ="""9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCAmelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) self.assertTrue(os.path.isfile(UpperCAmelCase_ ) ) # File is cached at the same place the second time. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Using a specific revision to test the full commit hash. lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''9b8c223''' ) self.assertEqual(UpperCAmelCase_ , os.path.join(UpperCAmelCase_ , '''snapshots''' , UpperCAmelCase_ , UpperCAmelCase_ ) ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): lowerCAmelCase = cached_file('''tiny-random-bert''' , UpperCAmelCase_ ) with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , UpperCAmelCase_ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) def __snake_case ( self ): with self.assertRaisesRegex(UpperCAmelCase_ , '''does not appear to have a file named''' ): lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' ) with open(os.path.join(UpperCAmelCase_ , '''refs''' , '''main''' ) ) as f: lowerCAmelCase = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCAmelCase_ , '''.no_exist''' , UpperCAmelCase_ , '''conf''' ) ) ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , local_files_only=UpperCAmelCase_ , _raise_exceptions_for_missing_entries=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) lowerCAmelCase = mock.Mock() lowerCAmelCase = 5_00 lowerCAmelCase = {} lowerCAmelCase = HTTPError lowerCAmelCase = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCAmelCase_ ) as mock_head: lowerCAmelCase = cached_file(UpperCAmelCase_ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCAmelCase_ ) self.assertIsNone(UpperCAmelCase_ ) # This check we did call the fake head request mock_head.assert_called() def __snake_case ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCAmelCase_ ) ) def __snake_case ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCAmelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCAmelCase_ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ , revision='''ahaha''' ) lowerCAmelCase = get_file_from_repo('''bert-base-cased''' , UpperCAmelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. lowerCAmelCase = json.loads(open(UpperCAmelCase_ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_68 ) def __snake_case ( self ): with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase = Path(UpperCAmelCase_ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCAmelCase_ , '''a.txt''' ) , str(UpperCAmelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCAmelCase_ , '''b.txt''' ) )
33
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } UpperCAmelCase_ =[ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models lowerCAmelCase = '''lm_head''' lowerCAmelCase = getattr(_snake_case , _snake_case ) if weight_type is not None: lowerCAmelCase = getattr(_snake_case , _snake_case ).shape else: lowerCAmelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowerCAmelCase = value elif weight_type == "weight_g": lowerCAmelCase = value elif weight_type == "weight_v": lowerCAmelCase = value elif weight_type == "bias": lowerCAmelCase = value else: lowerCAmelCase = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [] lowerCAmelCase = fairseq_model.state_dict() lowerCAmelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): lowerCAmelCase = False if "conv_layers" in name: load_conv_layer( _snake_case , _snake_case , _snake_case , _snake_case , hf_model.config.feat_extract_norm == '''group''' , ) lowerCAmelCase = True else: for key, mapped_key in MAPPING.items(): lowerCAmelCase = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: lowerCAmelCase = True if "*" in mapped_key: lowerCAmelCase = name.split(_snake_case )[0].split('''.''' )[-2] lowerCAmelCase = mapped_key.replace('''*''' , _snake_case ) if "weight_g" in name: lowerCAmelCase = '''weight_g''' elif "weight_v" in name: lowerCAmelCase = '''weight_v''' elif "bias" in name: lowerCAmelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj lowerCAmelCase = '''weight''' else: lowerCAmelCase = None set_recursively(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) continue if not is_used: unused_weights.append(_snake_case ) logger.warning(F"""Unused weights: {unused_weights}""" ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase = full_name.split('''conv_layers.''' )[-1] lowerCAmelCase = name.split('''.''' ) lowerCAmelCase = int(items[0] ) lowerCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowerCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowerCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) lowerCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) lowerCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_snake_case ) @torch.no_grad() def UpperCAmelCase ( _snake_case , _snake_case , _snake_case=None , _snake_case=None , _snake_case=True ): if config_path is not None: lowerCAmelCase = UniSpeechConfig.from_pretrained(_snake_case ) else: lowerCAmelCase = UniSpeechConfig() if is_finetuned: if dict_path: lowerCAmelCase = Dictionary.load_from_json(_snake_case ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCAmelCase = target_dict.pad_index lowerCAmelCase = target_dict.bos_index lowerCAmelCase = target_dict.eos_index lowerCAmelCase = len(target_dict.symbols ) lowerCAmelCase = os.path.join(_snake_case , '''vocab.json''' ) if not os.path.isdir(_snake_case ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(_snake_case ) ) return os.makedirs(_snake_case , exist_ok=_snake_case ) lowerCAmelCase = target_dict.indices # fairseq has the <pad> and <s> switched lowerCAmelCase = 42 lowerCAmelCase = 43 with open(_snake_case , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(_snake_case , _snake_case ) lowerCAmelCase = WavaVecaPhonemeCTCTokenizer( _snake_case , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=_snake_case , ) lowerCAmelCase = True if config.feat_extract_norm == '''layer''' else False lowerCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_snake_case , return_attention_mask=_snake_case , ) lowerCAmelCase = WavaVecaProcessor(feature_extractor=_snake_case , tokenizer=_snake_case ) processor.save_pretrained(_snake_case ) lowerCAmelCase = UniSpeechForCTC(_snake_case ) else: lowerCAmelCase = UniSpeechForPreTraining(_snake_case ) if is_finetuned: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) lowerCAmelCase = model[0].eval() recursively_load_weights(_snake_case , _snake_case , _snake_case ) hf_unispeech.save_pretrained(_snake_case ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) UpperCAmelCase_ =parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
716
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = None , UpperCAmelCase_ = False , UpperCAmelCase_ = None , UpperCAmelCase_ = True , UpperCAmelCase_ = "arrow" , **UpperCAmelCase_ , ): super().__init__( split=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , keep_in_memory=UpperCAmelCase_ , streaming=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = load_from_cache_file lowerCAmelCase = file_format lowerCAmelCase = Spark( df=UpperCAmelCase_ , features=UpperCAmelCase_ , cache_dir=UpperCAmelCase_ , working_dir=UpperCAmelCase_ , **UpperCAmelCase_ , ) def __snake_case ( self ): if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
33
0
from typing import TYPE_CHECKING from ...utils import _LazyModule UpperCAmelCase_ ={"""processing_wav2vec2_with_lm""": ["""Wav2Vec2ProcessorWithLM"""]} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
717
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def UpperCAmelCase ( _snake_case = 3 ): if isinstance(_snake_case , _snake_case ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(_snake_case ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) lowerCAmelCase = QuantumRegister(_snake_case , '''qr''' ) lowerCAmelCase = ClassicalRegister(_snake_case , '''cr''' ) lowerCAmelCase = QuantumCircuit(_snake_case , _snake_case ) lowerCAmelCase = number_of_qubits for i in range(_snake_case ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(_snake_case ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , _snake_case , _snake_case ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(_snake_case , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(_snake_case , _snake_case ) # simulate with 10000 shots lowerCAmelCase = Aer.get_backend('''qasm_simulator''' ) lowerCAmelCase = execute(_snake_case , _snake_case , shots=10000 ) return job.result().get_counts(_snake_case ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
33
0
from math import ceil def UpperCAmelCase ( _snake_case = 1001 ): lowerCAmelCase = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): lowerCAmelCase = 2 * i + 1 lowerCAmelCase = 2 * i lowerCAmelCase = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: UpperCAmelCase_ =int(sys.argv[1]) print(solution(n)) except ValueError: print("""Invalid entry - please enter a number""")
718
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class __UpperCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __a : Any =1 @register_to_config def __init__( self , UpperCAmelCase_=20_00 , UpperCAmelCase_=0.1 , UpperCAmelCase_=20 , UpperCAmelCase_=1E-3 ): lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = torch.linspace(1 , self.config.sampling_eps , UpperCAmelCase_ , device=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowerCAmelCase = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowerCAmelCase = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowerCAmelCase = std.flatten() while len(std.shape ) < len(score.shape ): lowerCAmelCase = std.unsqueeze(-1 ) lowerCAmelCase = -score / std # compute lowerCAmelCase = -1.0 / len(self.timesteps ) lowerCAmelCase = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowerCAmelCase = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowerCAmelCase = beta_t.unsqueeze(-1 ) lowerCAmelCase = -0.5 * beta_t * x lowerCAmelCase = torch.sqrt(UpperCAmelCase_ ) lowerCAmelCase = drift - diffusion**2 * score lowerCAmelCase = x + drift * dt # add noise lowerCAmelCase = randn_tensor(x.shape , layout=x.layout , generator=UpperCAmelCase_ , device=x.device , dtype=x.dtype ) lowerCAmelCase = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ): return self.config.num_train_timesteps
33
0
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py UpperCAmelCase_ ="""src/transformers""" # This is to make sure the transformers module imported is the one in the repo. UpperCAmelCase_ =direct_transformers_import(PATH_TO_TRANSFORMERS) UpperCAmelCase_ =transformers.models.auto.configuration_auto.CONFIG_MAPPING UpperCAmelCase_ ={ # used to compute the property `self.chunk_length` """EncodecConfig""": ["""overlap"""], # used as `self.bert_model = BertModel(config, ...)` """DPRConfig""": True, # not used in modeling files, but it's an important information """FSMTConfig""": ["""langs"""], # used internally in the configuration class file """GPTNeoConfig""": ["""attention_types"""], # used internally in the configuration class file """EsmConfig""": ["""is_folding_model"""], # used during training (despite we don't have training script for these models yet) """Mask2FormerConfig""": ["""ignore_value"""], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) """OneFormerConfig""": ["""ignore_value""", """norm"""], # used during preprocessing and collation, see `collating_graphormer.py` """GraphormerConfig""": ["""spatial_pos_max"""], # used internally in the configuration class file """T5Config""": ["""feed_forward_proj"""], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally """MT5Config""": ["""feed_forward_proj""", """tokenizer_class"""], """UMT5Config""": ["""feed_forward_proj""", """tokenizer_class"""], # used internally in the configuration class file """LongT5Config""": ["""feed_forward_proj"""], # used internally in the configuration class file """SwitchTransformersConfig""": ["""feed_forward_proj"""], # having default values other than `1e-5` - we can't fix them without breaking """BioGptConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """GLPNConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """SegformerConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """CvtConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """PerceiverConfig""": ["""layer_norm_eps"""], # used internally to calculate the feature size """InformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate the feature size """TimeSeriesTransformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate the feature size """AutoformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate `mlp_dim` """SamVisionConfig""": ["""mlp_ratio"""], # For (head) training, but so far not implemented """ClapAudioConfig""": ["""num_classes"""], # Not used, but providing useful information to users """SpeechT5HifiGanConfig""": ["""sampling_rate"""], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { """CLIPSegConfig""": True, """DeformableDetrConfig""": True, """DetaConfig""": True, """DinatConfig""": True, """DonutSwinConfig""": True, """EfficientFormerConfig""": True, """FSMTConfig""": True, """JukeboxConfig""": True, """LayoutLMv2Config""": True, """MaskFormerSwinConfig""": True, """MT5Config""": True, """NatConfig""": True, """OneFormerConfig""": True, """PerceiverConfig""": True, """RagConfig""": True, """SpeechT5Config""": True, """SwinConfig""": True, """Swin2SRConfig""": True, """Swinv2Config""": True, """SwitchTransformersConfig""": True, """TableTransformerConfig""": True, """TapasConfig""": True, """TransfoXLConfig""": True, """UniSpeechConfig""": True, """UniSpeechSatConfig""": True, """WavLMConfig""": True, """WhisperConfig""": True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) """JukeboxPriorConfig""": True, # TODO: @Younes (for `is_decoder`) """Pix2StructTextConfig""": True, } ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( F"""config.{attribute}""" in modeling_source or F"""getattr(config, \"{attribute}\"""" in modeling_source or F"""getattr(self.config, \"{attribute}\"""" in modeling_source ): lowerCAmelCase = True # Deal with multi-line cases elif ( re.search( RF"""getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*\"{attribute}\"""" , _snake_case , ) is not None ): lowerCAmelCase = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: lowerCAmelCase = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files lowerCAmelCase = [ '''bos_index''', '''eos_index''', '''pad_index''', '''unk_index''', '''mask_index''', '''image_size''', '''use_cache''', '''out_features''', '''out_indices''', ] lowerCAmelCase = ['''encoder_no_repeat_ngram_size'''] # Special cases to be allowed lowerCAmelCase = True if not attribute_used: lowerCAmelCase = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: lowerCAmelCase = True elif attribute in ["tie_word_embeddings"] and default_value is False: lowerCAmelCase = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: lowerCAmelCase = True elif attribute.endswith('''_token_id''' ): lowerCAmelCase = True # configuration class specific cases if not case_allowed: lowerCAmelCase = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , [] ) lowerCAmelCase = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def UpperCAmelCase ( _snake_case ): lowerCAmelCase = dict(inspect.signature(config_class.__init__ ).parameters ) lowerCAmelCase = [x for x in list(signature.keys() ) if x not in ['''self''', '''kwargs''']] lowerCAmelCase = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass lowerCAmelCase = {} if len(config_class.attribute_map ) > 0: lowerCAmelCase = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files lowerCAmelCase = inspect.getsourcefile(_snake_case ) lowerCAmelCase = os.path.dirname(_snake_case ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. lowerCAmelCase = [os.path.join(_snake_case , _snake_case ) for fn in os.listdir(_snake_case ) if fn.startswith('''modeling_''' )] # Get the source code strings lowerCAmelCase = [] for path in modeling_paths: if os.path.isfile(_snake_case ): with open(_snake_case ) as fp: modeling_sources.append(fp.read() ) lowerCAmelCase = [] for config_param, default_value in zip(_snake_case , _snake_case ): # `attributes` here is all the variant names for `config_param` lowerCAmelCase = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(_snake_case , _snake_case , _snake_case , _snake_case ): unused_attributes.append(attributes[0] ) return sorted(_snake_case ) def UpperCAmelCase ( ): lowerCAmelCase = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) lowerCAmelCase = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ) , lambda _snake_case : inspect.isclass(_snake_case ) and issubclass(_snake_case , _snake_case ) and inspect.getmodule(_snake_case ) == inspect.getmodule(_config_class ) , ) ] for config_class in config_classes_in_module: lowerCAmelCase = check_config_attributes_being_used(_snake_case ) if len(_snake_case ) > 0: lowerCAmelCase = unused_attributes if len(_snake_case ) > 0: lowerCAmelCase = '''The following configuration classes contain unused attributes in the corresponding modeling files:\n''' for name, attributes in configs_with_unused_attributes.items(): error += F"""{name}: {attributes}\n""" raise ValueError(_snake_case ) if __name__ == "__main__": check_config_attributes()
719
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class __UpperCamelCase ( yaml.SafeLoader ): '''simple docstring''' def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCAmelCase = [tuple(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else key for key in keys] lowerCAmelCase = Counter(UpperCAmelCase_ ) lowerCAmelCase = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): lowerCAmelCase = super().construct_mapping(UpperCAmelCase_ , deep=UpperCAmelCase_ ) self._check_no_duplicates_on_constructed_node(UpperCAmelCase_ ) return mapping def UpperCAmelCase ( _snake_case ): lowerCAmelCase = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCAmelCase = full_content[1:].index('''---''' ) + 1 lowerCAmelCase = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(_snake_case ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Any ={"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def __snake_case ( cls , UpperCAmelCase_ ): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(UpperCAmelCase_ ) else: return cls() def __snake_case ( self , UpperCAmelCase_ ): if path.exists(): with open(UpperCAmelCase_ , encoding='''utf-8''' ) as readme_file: lowerCAmelCase = readme_file.read() else: lowerCAmelCase = None lowerCAmelCase = self._to_readme(UpperCAmelCase_ ) with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ = None ): if readme_content is not None: lowerCAmelCase , lowerCAmelCase = _split_yaml_from_readme(UpperCAmelCase_ ) lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: lowerCAmelCase = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def __snake_case ( cls , UpperCAmelCase_ ): lowerCAmelCase = yaml.load(UpperCAmelCase_ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCAmelCase = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCAmelCase_ ) def __snake_case ( self ): return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCAmelCase_ , allow_unicode=UpperCAmelCase_ , encoding='''utf-8''' , ).decode('''utf-8''' ) UpperCAmelCase_ ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser UpperCAmelCase_ =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") UpperCAmelCase_ =ap.parse_args() UpperCAmelCase_ =Path(args.readme_filepath) UpperCAmelCase_ =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
33
0
class __UpperCamelCase : '''simple docstring''' def __init__( self ): lowerCAmelCase = {} # Mapping from char to TrieNode lowerCAmelCase = False def __snake_case ( self , UpperCAmelCase_ ): for word in words: self.insert(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self for char in word: if char not in curr.nodes: lowerCAmelCase = TrieNode() lowerCAmelCase = curr.nodes[char] lowerCAmelCase = True def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self for char in word: if char not in curr.nodes: return False lowerCAmelCase = curr.nodes[char] return curr.is_leaf def __snake_case ( self , UpperCAmelCase_ ): def _delete(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) -> bool: if index == len(UpperCAmelCase_ ): # If word does not exist if not curr.is_leaf: return False lowerCAmelCase = False return len(curr.nodes ) == 0 lowerCAmelCase = word[index] lowerCAmelCase = curr.nodes.get(UpperCAmelCase_ ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted lowerCAmelCase = _delete(UpperCAmelCase_ , UpperCAmelCase_ , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , UpperCAmelCase_ , 0 ) def UpperCAmelCase ( _snake_case , _snake_case ): if node.is_leaf: print(_snake_case , end=''' ''' ) for key, value in node.nodes.items(): print_words(_snake_case , word + key ) def UpperCAmelCase ( ): lowerCAmelCase = '''banana bananas bandana band apple all beast'''.split() lowerCAmelCase = TrieNode() root.insert_many(_snake_case ) # print_words(root, "") assert all(root.find(_snake_case ) for word in words ) assert root.find('''banana''' ) assert not root.find('''bandanas''' ) assert not root.find('''apps''' ) assert root.find('''apple''' ) assert root.find('''all''' ) root.delete('''all''' ) assert not root.find('''all''' ) root.delete('''banana''' ) assert not root.find('''banana''' ) assert root.find('''bananas''' ) return True def UpperCAmelCase ( _snake_case , _snake_case ): print(str(_snake_case ) , '''works!''' if passes else '''doesn\'t work :(''' ) def UpperCAmelCase ( ): assert test_trie() def UpperCAmelCase ( ): print_results('''Testing trie functionality''' , test_trie() ) if __name__ == "__main__": main()
720
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __a : str =MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = hf_hub_download( repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = VideoClassificationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ , top_k=2 ) lowerCAmelCase = [ example_video_filepath, '''https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4''', ] return video_classifier, examples def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): for example in examples: lowerCAmelCase = video_classifier(UpperCAmelCase_ ) self.assertEqual( UpperCAmelCase_ , [ {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, {'''score''': ANY(UpperCAmelCase_ ), '''label''': ANY(UpperCAmelCase_ )}, ] , ) @require_torch def __snake_case ( self ): lowerCAmelCase = '''hf-internal-testing/tiny-random-VideoMAEForVideoClassification''' lowerCAmelCase = VideoMAEFeatureExtractor( size={'''shortest_edge''': 10} , crop_size={'''height''': 10, '''width''': 10} ) lowerCAmelCase = pipeline( '''video-classification''' , model=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='''nateraw/video-demo''' , filename='''archery.mp4''' , repo_type='''dataset''' ) lowerCAmelCase = video_classifier(UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [ [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], [{'''score''': 0.5199, '''label''': '''LABEL_0'''}, {'''score''': 0.4801, '''label''': '''LABEL_1'''}], ] , ) @require_tf def __snake_case ( self ): pass
33
0
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig UpperCAmelCase_ =logging.get_logger(__name__) # General docstring UpperCAmelCase_ ="""PoolFormerConfig""" # Base docstring UpperCAmelCase_ ="""sail/poolformer_s12""" UpperCAmelCase_ =[1, 512, 7, 7] # Image classification docstring UpperCAmelCase_ ="""sail/poolformer_s12""" UpperCAmelCase_ ="""tabby, tabby cat""" UpperCAmelCase_ =[ """sail/poolformer_s12""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def UpperCAmelCase ( _snake_case , _snake_case = 0.0 , _snake_case = False ): if drop_prob == 0.0 or not training: return input lowerCAmelCase = 1 - drop_prob lowerCAmelCase = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets lowerCAmelCase = keep_prob + torch.rand(_snake_case , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize lowerCAmelCase = input.div(_snake_case ) * random_tensor return output class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ = None ): super().__init__() lowerCAmelCase = drop_prob def __snake_case ( self , UpperCAmelCase_ ): return drop_path(UpperCAmelCase_ , self.drop_prob , self.training ) def __snake_case ( self ): return "p={}".format(self.drop_prob ) class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None ): super().__init__() lowerCAmelCase = patch_size if isinstance(UpperCAmelCase_ , collections.abc.Iterable ) else (patch_size, patch_size) lowerCAmelCase = stride if isinstance(UpperCAmelCase_ , collections.abc.Iterable ) else (stride, stride) lowerCAmelCase = padding if isinstance(UpperCAmelCase_ , collections.abc.Iterable ) else (padding, padding) lowerCAmelCase = nn.Convad(UpperCAmelCase_ , UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ ) lowerCAmelCase = norm_layer(UpperCAmelCase_ ) if norm_layer else nn.Identity() def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.projection(UpperCAmelCase_ ) lowerCAmelCase = self.norm(UpperCAmelCase_ ) return embeddings class __UpperCamelCase ( nn.GroupNorm ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , **UpperCAmelCase_ ): super().__init__(1 , UpperCAmelCase_ , **UpperCAmelCase_ ) class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): super().__init__() lowerCAmelCase = nn.AvgPoolad(UpperCAmelCase_ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): return self.pool(UpperCAmelCase_ ) - hidden_states class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): super().__init__() lowerCAmelCase = nn.Convad(UpperCAmelCase_ , UpperCAmelCase_ , 1 ) lowerCAmelCase = nn.Convad(UpperCAmelCase_ , UpperCAmelCase_ , 1 ) lowerCAmelCase = PoolFormerDropPath(UpperCAmelCase_ ) if isinstance(config.hidden_act , UpperCAmelCase_ ): lowerCAmelCase = ACTaFN[config.hidden_act] else: lowerCAmelCase = config.hidden_act def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.conva(UpperCAmelCase_ ) lowerCAmelCase = self.act_fn(UpperCAmelCase_ ) lowerCAmelCase = self.drop(UpperCAmelCase_ ) lowerCAmelCase = self.conva(UpperCAmelCase_ ) lowerCAmelCase = self.drop(UpperCAmelCase_ ) return hidden_states class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): super().__init__() lowerCAmelCase = PoolFormerPooling(UpperCAmelCase_ ) lowerCAmelCase = PoolFormerOutput(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = PoolFormerGroupNorm(UpperCAmelCase_ ) lowerCAmelCase = PoolFormerGroupNorm(UpperCAmelCase_ ) # Useful for training neural nets lowerCAmelCase = PoolFormerDropPath(UpperCAmelCase_ ) if drop_path > 0.0 else nn.Identity() lowerCAmelCase = config.use_layer_scale if config.use_layer_scale: lowerCAmelCase = nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCAmelCase_) ) , requires_grad=UpperCAmelCase_ ) lowerCAmelCase = nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCAmelCase_) ) , requires_grad=UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): if self.use_layer_scale: lowerCAmelCase = self.pooling(self.before_norm(UpperCAmelCase_ ) ) lowerCAmelCase = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection lowerCAmelCase = hidden_states + self.drop_path(UpperCAmelCase_ ) lowerCAmelCase = () lowerCAmelCase = self.output(self.after_norm(UpperCAmelCase_ ) ) lowerCAmelCase = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection lowerCAmelCase = hidden_states + self.drop_path(UpperCAmelCase_ ) lowerCAmelCase = (output,) + outputs return outputs else: lowerCAmelCase = self.drop_path(self.pooling(self.before_norm(UpperCAmelCase_ ) ) ) # First residual connection lowerCAmelCase = pooling_output + hidden_states lowerCAmelCase = () # Second residual connection inside the PoolFormerOutput block lowerCAmelCase = self.drop_path(self.output(self.after_norm(UpperCAmelCase_ ) ) ) lowerCAmelCase = hidden_states + layer_output lowerCAmelCase = (output,) + outputs return outputs class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): super().__init__() lowerCAmelCase = config # stochastic depth decay rule lowerCAmelCase = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings lowerCAmelCase = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) lowerCAmelCase = nn.ModuleList(UpperCAmelCase_ ) # Transformer blocks lowerCAmelCase = [] lowerCAmelCase = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers lowerCAmelCase = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCAmelCase_ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCAmelCase_ ) ) lowerCAmelCase = nn.ModuleList(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False , UpperCAmelCase_=True ): lowerCAmelCase = () if output_hidden_states else None lowerCAmelCase = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): lowerCAmelCase , lowerCAmelCase = layers # Get patch embeddings from hidden_states lowerCAmelCase = embedding_layer(UpperCAmelCase_ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCAmelCase_ ): lowerCAmelCase = blk(UpperCAmelCase_ ) lowerCAmelCase = layer_outputs[0] if output_hidden_states: lowerCAmelCase = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Tuple =PoolFormerConfig __a : List[Any] ="""poolformer""" __a : Dict ="""pixel_values""" __a : int =True def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase = value UpperCAmelCase_ =R""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ UpperCAmelCase_ =R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`PoolFormerImageProcessor.__call__`] for details. """ @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , __UpperCAmelCase , ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): super().__init__(UpperCAmelCase_ ) lowerCAmelCase = config lowerCAmelCase = PoolFormerEncoder(UpperCAmelCase_ ) # Initialize weights and apply final processing self.post_init() def __snake_case ( self ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __snake_case ( self , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = None , ): lowerCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowerCAmelCase = self.encoder( UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ , ) lowerCAmelCase = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCAmelCase_ , hidden_states=encoder_outputs.hidden_states , ) class __UpperCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): super().__init__() lowerCAmelCase = nn.Linear(config.hidden_size , config.hidden_size ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.dense(UpperCAmelCase_ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , __UpperCAmelCase , ) class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): super().__init__(UpperCAmelCase_ ) lowerCAmelCase = config.num_labels lowerCAmelCase = PoolFormerModel(UpperCAmelCase_ ) # Final norm lowerCAmelCase = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head lowerCAmelCase = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __snake_case ( self , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = None , UpperCAmelCase_ = None , ): lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase = self.poolformer( UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ , ) lowerCAmelCase = outputs[0] lowerCAmelCase = self.classifier(self.norm(UpperCAmelCase_ ).mean([-2, -1] ) ) lowerCAmelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCAmelCase = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCAmelCase = '''single_label_classification''' else: lowerCAmelCase = '''multi_label_classification''' if self.config.problem_type == "regression": lowerCAmelCase = MSELoss() if self.num_labels == 1: lowerCAmelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowerCAmelCase = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": lowerCAmelCase = CrossEntropyLoss() lowerCAmelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowerCAmelCase = BCEWithLogitsLoss() lowerCAmelCase = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: lowerCAmelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states )
721
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Any ="""hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __snake_case ( self , UpperCAmelCase_=0 ): lowerCAmelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCAmelCase_ ) ) lowerCAmelCase = np.random.RandomState(UpperCAmelCase_ ) lowerCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # warmup pass to apply optimizations lowerCAmelCase = pipe(**self.get_dummy_inputs() ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __snake_case ( self ): lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = self.get_dummy_inputs() lowerCAmelCase = pipe(**UpperCAmelCase_ ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowerCAmelCase = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @property def __snake_case ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __snake_case ( self ): lowerCAmelCase = ort.SessionOptions() lowerCAmelCase = False return options def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __snake_case ( self ): lowerCAmelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) lowerCAmelCase = init_image.resize((7_68, 5_12) ) lowerCAmelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) lowerCAmelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=UpperCAmelCase_ , safety_checker=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) lowerCAmelCase = '''A fantasy landscape, trending on artstation''' lowerCAmelCase = np.random.RandomState(0 ) lowerCAmelCase = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase_ , output_type='''np''' , ) lowerCAmelCase = output.images lowerCAmelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowerCAmelCase = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
33
0
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=7 , UpperCAmelCase_=3 , UpperCAmelCase_=18 , UpperCAmelCase_=30 , UpperCAmelCase_=4_00 , UpperCAmelCase_=True , UpperCAmelCase_=None , UpperCAmelCase_=True , ): lowerCAmelCase = size if size is not None else {'''height''': 18, '''width''': 18} lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = num_channels lowerCAmelCase = image_size lowerCAmelCase = min_resolution lowerCAmelCase = max_resolution lowerCAmelCase = do_resize lowerCAmelCase = size lowerCAmelCase = do_normalize def __snake_case ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866443634033203, 0.6618829369544983, 0.3891746401786804], [-0.6042559146881104, -0.02295008860528469, 0.5423797369003296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Optional[Any] =ImageGPTImageProcessor if is_vision_available() else None def __snake_case ( self ): lowerCAmelCase = ImageGPTImageProcessingTester(self ) @property def __snake_case ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self ): lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase_ , '''clusters''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''do_normalize''' ) ) def __snake_case ( self ): lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def __snake_case ( self ): lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) lowerCAmelCase = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCAmelCase_ , obj[key] ) ) else: self.assertEqual(obj[key] , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase = os.path.join(UpperCAmelCase_ , '''image_processor.json''' ) image_processor_first.to_json_file(UpperCAmelCase_ ) lowerCAmelCase = self.image_processing_class.from_json_file(UpperCAmelCase_ ).to_dict() lowerCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCAmelCase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = self.image_processing_class.from_pretrained(UpperCAmelCase_ ).to_dict() lowerCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCAmelCase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , UpperCAmelCase_ ) @unittest.skip('''ImageGPT requires clusters at initialization''' ) def __snake_case ( self ): pass def UpperCAmelCase ( ): lowerCAmelCase = load_dataset('''hf-internal-testing/fixtures_image_utils''' , split='''test''' ) lowerCAmelCase = Image.open(dataset[4]['''file'''] ) lowerCAmelCase = Image.open(dataset[5]['''file'''] ) lowerCAmelCase = [imagea, imagea] return images @require_vision @require_torch class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __snake_case ( self ): lowerCAmelCase = ImageGPTImageProcessor.from_pretrained('''openai/imagegpt-small''' ) lowerCAmelCase = prepare_images() # test non-batched lowerCAmelCase = image_processing(images[0] , return_tensors='''pt''' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 10_24) ) lowerCAmelCase = [3_06, 1_91, 1_91] self.assertEqual(encoding.input_ids[0, :3].tolist() , UpperCAmelCase_ ) # test batched lowerCAmelCase = image_processing(UpperCAmelCase_ , return_tensors='''pt''' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 10_24) ) lowerCAmelCase = [3_03, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , UpperCAmelCase_ )
700
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def UpperCAmelCase ( _snake_case ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('''/''' ) lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) lowerCAmelCase = torch.load(os.path.join(_snake_case , '''pytorch_model.bin''' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_snake_case , threshold=_snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = TopKBinarizer.apply(_snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase = ThresholdBinarizer.apply(_snake_case , _snake_case , _snake_case ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F"""{prefix_}mask_scores"""] lowerCAmelCase , lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_snake_case ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_snake_case ) , F"""bertarized_{os.path.basename(_snake_case )}""" ) if not os.path.isdir(_snake_case ): shutil.copytree(_snake_case , _snake_case ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(_snake_case , os.path.join(_snake_case , '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": UpperCAmelCase_ =argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) UpperCAmelCase_ =parser.parse_args() main(args)
33
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class __UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=7 , UpperCAmelCase_=3 , UpperCAmelCase_=30 , UpperCAmelCase_=4_00 , UpperCAmelCase_=True , UpperCAmelCase_=None , UpperCAmelCase_=True , UpperCAmelCase_=[0.5, 0.5, 0.5] , UpperCAmelCase_=[0.5, 0.5, 0.5] , UpperCAmelCase_=True , UpperCAmelCase_=1 / 2_55 , UpperCAmelCase_=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowerCAmelCase = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 13_33} lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = num_channels lowerCAmelCase = min_resolution lowerCAmelCase = max_resolution lowerCAmelCase = do_resize lowerCAmelCase = size lowerCAmelCase = do_normalize lowerCAmelCase = image_mean lowerCAmelCase = image_std lowerCAmelCase = do_rescale lowerCAmelCase = rescale_factor lowerCAmelCase = do_pad def __snake_case ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_=False ): if not batched: lowerCAmelCase = image_inputs[0] if isinstance(UpperCAmelCase_ , Image.Image ): lowerCAmelCase , lowerCAmelCase = image.size else: lowerCAmelCase , lowerCAmelCase = image.shape[1], image.shape[2] if w < h: lowerCAmelCase = int(self.size['''shortest_edge'''] * h / w ) lowerCAmelCase = self.size['''shortest_edge'''] elif w > h: lowerCAmelCase = self.size['''shortest_edge'''] lowerCAmelCase = int(self.size['''shortest_edge'''] * w / h ) else: lowerCAmelCase = self.size['''shortest_edge'''] lowerCAmelCase = self.size['''shortest_edge'''] else: lowerCAmelCase = [] for image in image_inputs: lowerCAmelCase , lowerCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCAmelCase = max(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : item[0] )[0] lowerCAmelCase = max(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __UpperCamelCase ( __UpperCAmelCase , unittest.TestCase ): '''simple docstring''' __a : Union[str, Any] =ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self ): lowerCAmelCase = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self ): lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase_ , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''image_std''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase_ , '''size''' ) ) def __snake_case ( self ): lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 13_33} ) self.assertEqual(image_processor.do_pad , UpperCAmelCase_ ) lowerCAmelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=UpperCAmelCase_ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , UpperCAmelCase_ ) def __snake_case ( self ): pass def __snake_case ( self ): # Initialize image_processing lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , Image.Image ) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ , batched=UpperCAmelCase_ ) lowerCAmelCase = image_processing(UpperCAmelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self ): # Initialize image_processing lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , numpify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase = image_processing(UpperCAmelCase_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ , batched=UpperCAmelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self ): # Initialize image_processing lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , torchify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) # Test not batched input lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase = image_processing(UpperCAmelCase_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase , lowerCAmelCase = self.image_processor_tester.get_expected_values(UpperCAmelCase_ , batched=UpperCAmelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self ): # prepare image and target lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: lowerCAmelCase = json.loads(f.read() ) lowerCAmelCase = {'''image_id''': 3_97_69, '''annotations''': target} # encode them lowerCAmelCase = ConditionalDetrImageProcessor.from_pretrained('''microsoft/conditional-detr-resnet-50''' ) lowerCAmelCase = image_processing(images=UpperCAmelCase_ , annotations=UpperCAmelCase_ , return_tensors='''pt''' ) # verify pixel values lowerCAmelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) # verify area lowerCAmelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase_ ) ) # verify boxes lowerCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase_ , atol=1E-3 ) ) # verify image_id lowerCAmelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase_ ) ) # verify is_crowd lowerCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase_ ) ) # verify class_labels lowerCAmelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase_ ) ) # verify orig_size lowerCAmelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase_ ) ) # verify size lowerCAmelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase_ ) ) @slow def __snake_case ( self ): # prepare image, target and masks_path lowerCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: lowerCAmelCase = json.loads(f.read() ) lowerCAmelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_97_69, '''segments_info''': target} lowerCAmelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them lowerCAmelCase = ConditionalDetrImageProcessor(format='''coco_panoptic''' ) lowerCAmelCase = image_processing(images=UpperCAmelCase_ , annotations=UpperCAmelCase_ , masks_path=UpperCAmelCase_ , return_tensors='''pt''' ) # verify pixel values lowerCAmelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) # verify area lowerCAmelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase_ ) ) # verify boxes lowerCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase_ ) lowerCAmelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase_ , atol=1E-3 ) ) # verify image_id lowerCAmelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase_ ) ) # verify is_crowd lowerCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase_ ) ) # verify class_labels lowerCAmelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase_ ) ) # verify masks lowerCAmelCase = 82_28_73 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , UpperCAmelCase_ ) # verify orig_size lowerCAmelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase_ ) ) # verify size lowerCAmelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase_ ) )
701
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ =logging.get_logger(__name__) UpperCAmelCase_ ={ """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } UpperCAmelCase_ ={ """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } UpperCAmelCase_ ={ """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def UpperCAmelCase ( _snake_case ): lowerCAmelCase = set() lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase = char lowerCAmelCase = set(_snake_case ) return pairs class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =VOCAB_FILES_NAMES __a : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP __a : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , **UpperCAmelCase_ , ): super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase = vocab_file lowerCAmelCase = merges_file lowerCAmelCase = {} lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 lowerCAmelCase = 3 self.add_from_file(UpperCAmelCase_ ) lowerCAmelCase = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase_ , encoding='''utf-8''' ) as merges_handle: lowerCAmelCase = merges_handle.read().split('''\n''' )[:-1] lowerCAmelCase = [tuple(merge.split()[:-1] ) for merge in merges] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None , UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __snake_case ( self ): return len(self.encoder ) def __snake_case ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self , UpperCAmelCase_ ): if token in self.cache: return self.cache[token] lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowerCAmelCase = get_pairs(UpperCAmelCase_ ) if not pairs: return token while True: lowerCAmelCase = min(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : self.bpe_ranks.get(UpperCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase = bigram lowerCAmelCase = [] lowerCAmelCase = 0 while i < len(UpperCAmelCase_ ): try: lowerCAmelCase = word.index(UpperCAmelCase_ , UpperCAmelCase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase = j if word[i] == first and i < len(UpperCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase = tuple(UpperCAmelCase_ ) lowerCAmelCase = new_word if len(UpperCAmelCase_ ) == 1: break else: lowerCAmelCase = get_pairs(UpperCAmelCase_ ) lowerCAmelCase = '''@@ '''.join(UpperCAmelCase_ ) lowerCAmelCase = word[:-4] lowerCAmelCase = word return word def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = [] lowerCAmelCase = re.findall(r'''\S+\n?''' , UpperCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __snake_case ( self , UpperCAmelCase_ ): return self.encoder.get(UpperCAmelCase_ , self.encoder.get(self.unk_token ) ) def __snake_case ( self , UpperCAmelCase_ ): return self.decoder.get(UpperCAmelCase_ , self.unk_token ) def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = ''' '''.join(UpperCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.merges_file , UpperCAmelCase_ ) return out_vocab_file, out_merge_file def __snake_case ( self , UpperCAmelCase_ ): if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(UpperCAmelCase_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCAmelCase = f.readlines() for lineTmp in lines: lowerCAmelCase = lineTmp.strip() lowerCAmelCase = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowerCAmelCase = line[:idx] lowerCAmelCase = len(self.encoder )
33
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ ={ """configuration_deberta""": ["""DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DebertaConfig""", """DebertaOnnxConfig"""], """tokenization_deberta""": ["""DebertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =["""DebertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """DebertaForMaskedLM""", """DebertaForQuestionAnswering""", """DebertaForSequenceClassification""", """DebertaForTokenClassification""", """DebertaModel""", """DebertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDebertaForMaskedLM""", """TFDebertaForQuestionAnswering""", """TFDebertaForSequenceClassification""", """TFDebertaForTokenClassification""", """TFDebertaModel""", """TFDebertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
702
from __future__ import annotations from typing import Generic, TypeVar UpperCAmelCase_ =TypeVar("""T""") class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self , UpperCAmelCase_ ): lowerCAmelCase = data lowerCAmelCase = self lowerCAmelCase = 0 class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # map from node name to the node object lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # create a new set with x as its member lowerCAmelCase = DisjointSetTreeNode(UpperCAmelCase_ ) def __snake_case ( self , UpperCAmelCase_ ): # find the set x belongs to (with path-compression) lowerCAmelCase = self.map[data] if elem_ref != elem_ref.parent: lowerCAmelCase = self.find_set(elem_ref.parent.data ) return elem_ref.parent def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # helper function for union operation if nodea.rank > nodea.rank: lowerCAmelCase = nodea else: lowerCAmelCase = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ ): # merge 2 disjoint sets self.link(self.find_set(UpperCAmelCase_ ) , self.find_set(UpperCAmelCase_ ) ) class __UpperCamelCase ( Generic[T] ): '''simple docstring''' def __init__( self ): # connections: map from the node to the neighbouring nodes (with weights) lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ ): # add a node ONLY if its not present in the graph if node not in self.connections: lowerCAmelCase = {} def __snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # add an edge with the given weight self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) lowerCAmelCase = weight lowerCAmelCase = weight def __snake_case ( self ): lowerCAmelCase = [] lowerCAmelCase = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda UpperCAmelCase_ : x[2] ) # creating the disjoint set lowerCAmelCase = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(UpperCAmelCase_ ) # MST generation lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = edges[index] index += 1 lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) lowerCAmelCase = disjoint_set.find_set(UpperCAmelCase_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) disjoint_set.union(UpperCAmelCase_ , UpperCAmelCase_ ) return graph
33
0
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' __a : Union[str, Any] =(EulerDiscreteScheduler,) __a : List[str] =1_0 def __snake_case ( self , **UpperCAmelCase_ ): lowerCAmelCase = { '''num_train_timesteps''': 11_00, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**UpperCAmelCase_ ) return config def __snake_case ( self ): for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=UpperCAmelCase_ ) def __snake_case ( self ): for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=UpperCAmelCase_ , beta_end=UpperCAmelCase_ ) def __snake_case ( self ): for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=UpperCAmelCase_ ) def __snake_case ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase_ ) def __snake_case ( self ): lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase = sample.to(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , generator=UpperCAmelCase_ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1E-2 assert abs(result_mean.item() - 0.0131 ) < 1E-3 def __snake_case ( self ): lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(prediction_type='''v_prediction''' ) lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase = sample.to(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , generator=UpperCAmelCase_ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 0.0002 ) < 1E-2 assert abs(result_mean.item() - 2.2_6_7_6E-0_6 ) < 1E-3 def __snake_case ( self ): lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase_ ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowerCAmelCase = sample.to(UpperCAmelCase_ ) for t in scheduler.timesteps: lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , generator=UpperCAmelCase_ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1E-2 assert abs(result_mean.item() - 0.0131 ) < 1E-3 def __snake_case ( self ): lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase_ , use_karras_sigmas=UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase_ ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowerCAmelCase = sample.to(UpperCAmelCase_ ) for t in scheduler.timesteps: lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = model(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , generator=UpperCAmelCase_ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 124.52299499511719 ) < 1E-2 assert abs(result_mean.item() - 0.16213932633399963 ) < 1E-3
703
def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations(_snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(_snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): def count_of_possible_combinations_with_dp_array( _snake_case , _snake_case ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase = sum( count_of_possible_combinations_with_dp_array(target - item , _snake_case ) for item in array ) lowerCAmelCase = answer return answer lowerCAmelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(_snake_case , _snake_case ) def UpperCAmelCase ( _snake_case , _snake_case , _snake_case ): lowerCAmelCase = [0] * (target + 1) lowerCAmelCase = 1 for i in range(1 , target + 1 ): for j in range(_snake_case ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ =3 UpperCAmelCase_ =5 UpperCAmelCase_ =[1, 2, 5] print(combination_sum_iv(n, array, target))
33
0
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def UpperCAmelCase ( _snake_case = "" ): lowerCAmelCase = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' lowerCAmelCase = BeautifulSoup(requests.get(_snake_case ).text , '''html.parser''' ) lowerCAmelCase = soup.find_all('''td''' , attrs='''titleColumn''' ) lowerCAmelCase = soup.find_all('''td''' , class_='''ratingColumn imdbRating''' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(_snake_case , _snake_case ) } def UpperCAmelCase ( _snake_case = "IMDb_Top_250_Movies.csv" ): lowerCAmelCase = get_imdb_top_aaa_movies() with open(_snake_case , '''w''' , newline='''''' ) as out_file: lowerCAmelCase = csv.writer(_snake_case ) writer.writerow(['''Movie title''', '''IMDb rating'''] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
704
import torch from diffusers import StableDiffusionPipeline UpperCAmelCase_ ="""path-to-your-trained-model""" UpperCAmelCase_ =StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") UpperCAmelCase_ ="""A photo of sks dog in a bucket""" UpperCAmelCase_ =pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
33
0
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def UpperCAmelCase ( _snake_case ): if not is_accelerate_available(): return method lowerCAmelCase = version.parse(accelerate.__version__ ).base_version if version.parse(_snake_case ) < version.parse('''0.17.0''' ): return method def wrapper(self , *_snake_case , **_snake_case ): if hasattr(self , '''_hf_hook''' ) and hasattr(self._hf_hook , '''pre_forward''' ): self._hf_hook.pre_forward(self ) return method(self , *_snake_case , **_snake_case ) return wrapper
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ ={ """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ =[ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCAmelCase_ =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
import numpy as np def UpperCAmelCase ( _snake_case ): return 1 / (1 + np.exp(-vector )) def UpperCAmelCase ( _snake_case ): return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
706
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __UpperCamelCase ( __UpperCAmelCase ): '''simple docstring''' def __snake_case ( self ): lowerCAmelCase = tempfile.mkdtemp() lowerCAmelCase = 8 # DPR tok lowerCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowerCAmelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] lowerCAmelCase = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) lowerCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase = {'''unk_token''': '''<unk>'''} lowerCAmelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase = os.path.join(UpperCAmelCase_ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCAmelCase_ ) ) def __snake_case ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __snake_case ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __snake_case ( self ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __snake_case ( self ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def __snake_case ( self , UpperCAmelCase_ ): lowerCAmelCase = self.get_dummy_dataset() lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: lowerCAmelCase = os.path.join(self.tmpdirname , '''dataset''' ) lowerCAmelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCAmelCase_ ) , ) return retriever def __snake_case ( self ): lowerCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) lowerCAmelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) lowerCAmelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(UpperCAmelCase_ , open(UpperCAmelCase_ , '''wb''' ) ) lowerCAmelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) lowerCAmelCase = RagRetriever( UpperCAmelCase_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: lowerCAmelCase = self.get_dummy_dataset() retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) def __snake_case ( self ): lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_legacy_index_retriever() lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=UpperCAmelCase_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCAmelCase_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCAmelCase_ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def __snake_case ( self ): lowerCAmelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCAmelCase_ ) lowerCAmelCase = RagRetriever.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever.retrieve(UpperCAmelCase_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): import torch lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) lowerCAmelCase = retriever( UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ , return_tensors='''pt''' , ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __snake_case ( self ): lowerCAmelCase = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase = 1 lowerCAmelCase = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCAmelCase_ ) retriever.set_ctx_encoder_tokenizer(UpperCAmelCase_ ) lowerCAmelCase = [[5, 7], [10, 11]] lowerCAmelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase = retriever(UpperCAmelCase_ , UpperCAmelCase_ , prefix=retriever.config.generator.prefix , n_docs=UpperCAmelCase_ ) self.assertEqual( len(UpperCAmelCase_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCAmelCase_ ) # check for doc token related keys in dictionary.
33
0