code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a__ : List[str] =logging.get_logger(__name__) a__ : Any ='''▁''' a__ : Dict ={'''vocab_file''': '''sentencepiece.bpe.model'''} a__ : List[Any] ={ '''vocab_file''': { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model''', } } a__ : List[Any] ={ '''facebook/xglm-564M''': 2_048, } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] =VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : List[Any] =PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Optional[Any] =["input_ids", "attention_mask"] def __init__( self : Dict , __A : List[Any] , __A : Dict="<s>" , __A : Any="</s>" , __A : Union[str, Any]="</s>" , __A : Optional[int]="<s>" , __A : Dict="<unk>" , __A : Optional[Any]="<pad>" , __A : Optional[Dict[str, Any]] = None , **__A : Union[str, Any] , ): __UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer __UpperCamelCase = 7 __UpperCamelCase = [f'''<madeupword{i}>''' for i in range(self.num_madeup_words )] __UpperCamelCase = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=__A , eos_token=__A , unk_token=__A , sep_token=__A , cls_token=__A , pad_token=__A , sp_model_kwargs=self.sp_model_kwargs , **__A , ) __UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__A ) ) __UpperCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __UpperCamelCase = 1 # Mimic fairseq token-to-id alignment for the first 4 token __UpperCamelCase = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} __UpperCamelCase = len(self.sp_model ) __UpperCamelCase = {f'''<madeupword{i}>''': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(__A ) __UpperCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Tuple ): __UpperCamelCase = self.__dict__.copy() __UpperCamelCase = None __UpperCamelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self : Any , __A : Optional[int] ): __UpperCamelCase = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __UpperCamelCase = {} __UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _lowerCamelCase ( self : Any , __A : List[int] , __A : Optional[List[int]] = None ): if token_ids_a is None: return [self.sep_token_id] + token_ids_a __UpperCamelCase = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _lowerCamelCase ( self : Any , __A : List[int] , __A : Optional[List[int]] = None , __A : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__A , token_ids_a=__A , already_has_special_tokens=__A ) if token_ids_a is None: return [1] + ([0] * len(__A )) return [1] + ([0] * len(__A )) + [1, 1] + ([0] * len(__A )) def _lowerCamelCase ( self : Dict , __A : List[int] , __A : Optional[List[int]] = None ): __UpperCamelCase = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _lowerCamelCase ( self : Optional[int] ): return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = {self.convert_ids_to_tokens(__A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowerCamelCase ( self : List[Any] , __A : str ): return self.sp_model.encode(__A , out_type=__A ) def _lowerCamelCase ( self : int , __A : Any ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __UpperCamelCase = self.sp_model.PieceToId(__A ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _lowerCamelCase ( self : List[str] , __A : List[str] ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _lowerCamelCase ( self : Dict , __A : int ): __UpperCamelCase = ''.join(__A ).replace(__A , ' ' ).strip() return out_string def _lowerCamelCase ( self : int , __A : str , __A : Optional[str] = None ): if not os.path.isdir(__A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __UpperCamelCase = os.path.join( __A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __A ) elif not os.path.isfile(self.vocab_file ): with open(__A , 'wb' ) as fi: __UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(__A ) return (out_vocab_file,)
53
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Any ) -> Optional[Any]: """simple docstring""" with open(__lowercase ) as metadata_file: __UpperCamelCase = json.load(__lowercase ) __UpperCamelCase = LukeConfig(use_entity_aware_attention=__lowercase , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __UpperCamelCase = torch.load(__lowercase , map_location='cpu' ) # Load the entity vocab file __UpperCamelCase = load_entity_vocab(__lowercase ) __UpperCamelCase = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __UpperCamelCase = AddedToken('<ent>' , lstrip=__lowercase , rstrip=__lowercase ) __UpperCamelCase = AddedToken('<ent2>' , lstrip=__lowercase , rstrip=__lowercase ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(__lowercase , __lowercase ) __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase ) # Initialize the embeddings of the special tokens __UpperCamelCase = state_dict['embeddings.word_embeddings.weight'] __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __UpperCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __UpperCamelCase = F'''encoder.layer.{layer_index}.attention.self.''' __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __UpperCamelCase = state_dict['entity_embeddings.entity_embeddings.weight'] __UpperCamelCase = entity_emb[entity_vocab['[MASK]']] __UpperCamelCase = LukeModel(config=__lowercase ).eval() __UpperCamelCase , __UpperCamelCase = model.load_state_dict(__lowercase , strict=__lowercase ) if not (len(__lowercase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {', '.join(__lowercase )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' F''' {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}''' ) # Check outputs __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase , task='entity_classification' ) __UpperCamelCase = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __UpperCamelCase = (39, 42) __UpperCamelCase = tokenizer(__lowercase , entity_spans=[span] , add_prefix_space=__lowercase , return_tensors='pt' ) __UpperCamelCase = model(**__lowercase ) # Verify word hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 42, 1024) ) __UpperCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base __UpperCamelCase = torch.Size((1, 42, 768) ) __UpperCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 1, 1024) ) __UpperCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base __UpperCamelCase = torch.Size((1, 1, 768) ) __UpperCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(__lowercase ) ) model.save_pretrained(__lowercase ) def lowercase__ ( __lowercase : Dict ) -> List[str]: """simple docstring""" __UpperCamelCase = {} with open(__lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(__lowercase ): __UpperCamelCase , __UpperCamelCase = line.rstrip().split('\t' ) __UpperCamelCase = index return entity_vocab if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : str =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
53
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor a__ : int =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , *__A : Optional[int] , **__A : str ): warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' , __A , ) super().__init__(*__A , **__A )
53
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case ( __lowerCamelCase ): """simple docstring""" def _lowerCamelCase ( self : Any ): __UpperCamelCase = tempfile.mkdtemp() __UpperCamelCase = 8 # DPR tok __UpperCamelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Optional[int] ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Union[str, Any] ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def _lowerCamelCase ( self : str ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _lowerCamelCase ( self : Any , __A : bool ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __UpperCamelCase = os.path.join(self.tmpdirname , 'dataset' ) __UpperCamelCase = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _lowerCamelCase ( self : int ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __UpperCamelCase = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __UpperCamelCase = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __UpperCamelCase = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__A , open(__A , 'wb' ) ) __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_legacy_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __A ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Optional[Any] ): import torch __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) __UpperCamelCase = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors='pt' , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dpr_ctx_encoder_tokenizer() __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __A ) # check for doc token related keys in dictionary.
53
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Tuple =logging.get_logger(__name__) a__ : str ={ '''RWKV/rwkv-4-169m-pile''': '''https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json''', '''RWKV/rwkv-4-430m-pile''': '''https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json''', '''RWKV/rwkv-4-1b5-pile''': '''https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json''', '''RWKV/rwkv-4-3b-pile''': '''https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json''', '''RWKV/rwkv-4-7b-pile''': '''https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json''', '''RWKV/rwkv-4-14b-pile''': '''https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json''', '''RWKV/rwkv-raven-1b5''': '''https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json''', '''RWKV/rwkv-raven-3b''': '''https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json''', '''RWKV/rwkv-raven-7b''': '''https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json''', '''RWKV/rwkv-raven-14b''': '''https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json''', } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int ="rwkv" SCREAMING_SNAKE_CASE_ : Optional[int] ={"max_position_embeddings": "context_length"} def __init__( self : str , __A : Optional[int]=5_0_2_7_7 , __A : Any=1_0_2_4 , __A : Optional[int]=4_0_9_6 , __A : Optional[Any]=3_2 , __A : str=None , __A : Dict=None , __A : List[str]=1e-5 , __A : Dict=0 , __A : List[Any]=0 , __A : Dict=6 , __A : Optional[Any]=False , __A : Union[str, Any]=True , **__A : Tuple , ): __UpperCamelCase = vocab_size __UpperCamelCase = context_length __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = attention_hidden_size if attention_hidden_size is not None else hidden_size __UpperCamelCase = intermediate_size if intermediate_size is not None else 4 * hidden_size __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = rescale_every __UpperCamelCase = use_cache __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id super().__init__( tie_word_embeddings=__A , bos_token_id=__A , eos_token_id=__A , **__A )
53
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[Any] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys a__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
1
'''simple docstring''' import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class snake_case : """simple docstring""" def __init__( self : int , __A : Optional[Any] ): if isinstance(__A , __A ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden __UpperCamelCase = deepcopy(__A ) elif os.path.exists(__A ): with io.open(__A , 'r' , encoding='utf-8' ) as f: __UpperCamelCase = json.load(__A ) else: try: __UpperCamelCase = baseaa.urlsafe_baadecode(__A ).decode('utf-8' ) __UpperCamelCase = json.loads(__A ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f'''Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}''' ) __UpperCamelCase = config self.set_stage_and_offload() def _lowerCamelCase ( self : str ): # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. __UpperCamelCase = self.get_value('zero_optimization.stage' , -1 ) # offload __UpperCamelCase = False if self.is_zeroa() or self.is_zeroa(): __UpperCamelCase = set(['cpu', 'nvme'] ) __UpperCamelCase = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: __UpperCamelCase = True def _lowerCamelCase ( self : Union[str, Any] , __A : List[str] ): __UpperCamelCase = self.config # find the config node of interest if it exists __UpperCamelCase = ds_key_long.split('.' ) __UpperCamelCase = nodes.pop() for node in nodes: __UpperCamelCase = config.get(__A ) if config is None: return None, ds_key return config, ds_key def _lowerCamelCase ( self : int , __A : Any , __A : Optional[int]=None ): __UpperCamelCase , __UpperCamelCase = self.find_config_node(__A ) if config is None: return default return config.get(__A , __A ) def _lowerCamelCase ( self : str , __A : Optional[int] , __A : Optional[int]=False ): __UpperCamelCase = self.config # find the config node of interest if it exists __UpperCamelCase = ds_key_long.split('.' ) for node in nodes: __UpperCamelCase = config __UpperCamelCase = config.get(__A ) if config is None: if must_exist: raise ValueError(f'''Can\'t find {ds_key_long} entry in the config: {self.config}''' ) else: return # if found remove it if parent_config is not None: parent_config.pop(__A ) def _lowerCamelCase ( self : Dict , __A : Tuple ): __UpperCamelCase = self.get_value(__A ) return False if value is None else bool(__A ) def _lowerCamelCase ( self : List[Any] , __A : Union[str, Any] ): __UpperCamelCase = self.get_value(__A ) return False if value is None else not bool(__A ) def _lowerCamelCase ( self : List[Any] ): return self._stage == 2 def _lowerCamelCase ( self : int ): return self._stage == 3 def _lowerCamelCase ( self : Any ): return self._offload class snake_case : """simple docstring""" def __init__( self : int , __A : List[Any] ): __UpperCamelCase = engine def _lowerCamelCase ( self : Tuple , __A : Optional[int] , **__A : List[Any] ): # runs backpropagation and handles mixed precision self.engine.backward(__A , **__A ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Union[str, Any] , __A : List[str] ): super().__init__(__A , device_placement=__A , scaler=__A ) __UpperCamelCase = hasattr(self.optimizer , 'overflow' ) def _lowerCamelCase ( self : str , __A : Dict=None ): pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _lowerCamelCase ( self : Dict ): pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _lowerCamelCase ( self : Optional[int] ): if self.__has_overflow__: return self.optimizer.overflow return False class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Any , __A : int , __A : Dict ): super().__init__(__A , __A ) def _lowerCamelCase ( self : Dict ): pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Tuple , __A : Optional[int]=0.001 , __A : str=0 , **__A : List[str] ): __UpperCamelCase = params __UpperCamelCase = lr __UpperCamelCase = weight_decay __UpperCamelCase = kwargs class snake_case : """simple docstring""" def __init__( self : str , __A : str , __A : List[Any]=None , __A : str=0 , **__A : Optional[int] ): __UpperCamelCase = optimizer __UpperCamelCase = total_num_steps __UpperCamelCase = warmup_num_steps __UpperCamelCase = kwargs
53
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a__ : Dict =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] =["pixel_values"] def __init__( self : int , __A : bool = True , __A : Dict[str, int] = None , __A : float = None , __A : PILImageResampling = PILImageResampling.BILINEAR , __A : bool = True , __A : Union[int, float] = 1 / 2_5_5 , __A : bool = True , __A : Optional[Union[float, List[float]]] = None , __A : Optional[Union[float, List[float]]] = None , **__A : Tuple , ): super().__init__(**__A ) __UpperCamelCase = size if size is not None else {'shortest_edge': 3_8_4} __UpperCamelCase = get_size_dict(__A , default_to_square=__A ) __UpperCamelCase = do_resize __UpperCamelCase = size # Default value set here for backwards compatibility where the value in config is None __UpperCamelCase = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 __UpperCamelCase = resample __UpperCamelCase = do_rescale __UpperCamelCase = rescale_factor __UpperCamelCase = do_normalize __UpperCamelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __UpperCamelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def _lowerCamelCase ( self : int , __A : np.ndarray , __A : Dict[str, int] , __A : float , __A : PILImageResampling = PILImageResampling.BICUBIC , __A : Optional[Union[str, ChannelDimension]] = None , **__A : str , ): __UpperCamelCase = get_size_dict(__A , default_to_square=__A ) if "shortest_edge" not in size: raise ValueError(f'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) __UpperCamelCase = size['shortest_edge'] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct __UpperCamelCase = int(shortest_edge / crop_pct ) __UpperCamelCase = get_resize_output_image_size(__A , size=__A , default_to_square=__A ) __UpperCamelCase = resize(image=__A , size=__A , resample=__A , data_format=__A , **__A ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__A , size=(shortest_edge, shortest_edge) , data_format=__A , **__A ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __A , size=(shortest_edge, shortest_edge) , resample=__A , data_format=__A , **__A ) def _lowerCamelCase ( self : int , __A : np.ndarray , __A : Union[int, float] , __A : Optional[Union[str, ChannelDimension]] = None , **__A : Optional[Any] , ): return rescale(__A , scale=__A , data_format=__A , **__A ) def _lowerCamelCase ( self : Tuple , __A : np.ndarray , __A : Union[float, List[float]] , __A : Union[float, List[float]] , __A : Optional[Union[str, ChannelDimension]] = None , **__A : Optional[int] , ): return normalize(__A , mean=__A , std=__A , data_format=__A , **__A ) def _lowerCamelCase ( self : str , __A : ImageInput , __A : bool = None , __A : Dict[str, int] = None , __A : float = None , __A : PILImageResampling = None , __A : bool = None , __A : float = None , __A : bool = None , __A : Optional[Union[float, List[float]]] = None , __A : Optional[Union[float, List[float]]] = None , __A : Optional[Union[str, TensorType]] = None , __A : ChannelDimension = ChannelDimension.FIRST , **__A : List[str] , ): __UpperCamelCase = do_resize if do_resize is not None else self.do_resize __UpperCamelCase = crop_pct if crop_pct is not None else self.crop_pct __UpperCamelCase = resample if resample is not None else self.resample __UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase = do_normalize if do_normalize is not None else self.do_normalize __UpperCamelCase = image_mean if image_mean is not None else self.image_mean __UpperCamelCase = image_std if image_std is not None else self.image_std __UpperCamelCase = size if size is not None else self.size __UpperCamelCase = get_size_dict(__A , default_to_square=__A ) __UpperCamelCase = make_list_of_images(__A ) if not valid_images(__A ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError('crop_pct must be specified if size < 384.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __UpperCamelCase = [to_numpy_array(__A ) for image in images] if do_resize: __UpperCamelCase = [self.resize(image=__A , size=__A , crop_pct=__A , resample=__A ) for image in images] if do_rescale: __UpperCamelCase = [self.rescale(image=__A , scale=__A ) for image in images] if do_normalize: __UpperCamelCase = [self.normalize(image=__A , mean=__A , std=__A ) for image in images] __UpperCamelCase = [to_channel_dimension_format(__A , __A ) for image in images] __UpperCamelCase = {'pixel_values': images} return BatchFeature(data=__A , tensor_type=__A )
53
'''simple docstring''' # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys a__ : Tuple ='''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
53
1
'''simple docstring''' from collections.abc import Sequence def lowercase__ ( __lowercase : Sequence[float] , __lowercase : bool = False ) -> float: """simple docstring""" if not arr: return 0 __UpperCamelCase = 0 if allow_empty_subarrays else float('-inf' ) __UpperCamelCase = 0.0 for num in arr: __UpperCamelCase = max(0 if allow_empty_subarrays else num , curr_sum + num ) __UpperCamelCase = max(__lowercase , __lowercase ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() a__ : Union[str, Any] =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(f'{max_subarray_sum(nums) = }')
53
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
1
'''simple docstring''' import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case : """simple docstring""" def __init__( self : Union[str, Any] , __A : str , __A : Tuple=1_3 , __A : Optional[int]=[3_0, 3_0] , __A : str=2 , __A : List[Any]=3 , __A : Dict=True , __A : Union[str, Any]=True , __A : Tuple=3_2 , __A : str=5 , __A : Dict=4 , __A : Optional[int]=3_7 , __A : Tuple="gelu" , __A : Tuple=0.1 , __A : List[str]=0.1 , __A : List[str]=1_0 , __A : Optional[int]=0.02 , __A : str=3 , __A : Dict=None , __A : List[str]=8 , __A : Any=1_0 , ): __UpperCamelCase = parent __UpperCamelCase = batch_size __UpperCamelCase = image_size __UpperCamelCase = patch_size __UpperCamelCase = num_channels __UpperCamelCase = is_training __UpperCamelCase = use_labels __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = type_sequence_label_size __UpperCamelCase = initializer_range __UpperCamelCase = num_labels __UpperCamelCase = scope __UpperCamelCase = n_targets __UpperCamelCase = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens __UpperCamelCase = (image_size[1] // patch_size) * (image_size[0] // patch_size) __UpperCamelCase = num_patches + 1 + self.num_detection_tokens def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) __UpperCamelCase = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) __UpperCamelCase = [] for i in range(self.batch_size ): __UpperCamelCase = {} __UpperCamelCase = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=__A ) __UpperCamelCase = torch.rand(self.n_targets , 4 , device=__A ) labels.append(__A ) __UpperCamelCase = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : Tuple ): return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__A , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def _lowerCamelCase ( self : Optional[Any] , __A : str , __A : Dict , __A : Dict ): __UpperCamelCase = YolosModel(config=__A ) model.to(__A ) model.eval() __UpperCamelCase = model(__A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def _lowerCamelCase ( self : Optional[Any] , __A : Union[str, Any] , __A : Dict , __A : Optional[Any] ): __UpperCamelCase = YolosForObjectDetection(__A ) model.to(__A ) model.eval() __UpperCamelCase = model(pixel_values=__A ) __UpperCamelCase = model(__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) __UpperCamelCase = model(pixel_values=__A , labels=__A ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = self.prepare_config_and_inputs() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = config_and_inputs __UpperCamelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class snake_case ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict =(YolosModel, YolosForObjectDetection) if is_torch_available() else () SCREAMING_SNAKE_CASE_ : str =( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Dict =False SCREAMING_SNAKE_CASE_ : Tuple =False SCREAMING_SNAKE_CASE_ : Optional[int] =False SCREAMING_SNAKE_CASE_ : Tuple =False def _lowerCamelCase ( self : Optional[Any] , __A : Optional[int] , __A : Dict , __A : str=False ): __UpperCamelCase = super()._prepare_for_class(__A , __A , return_labels=__A ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": __UpperCamelCase = [] for i in range(self.model_tester.batch_size ): __UpperCamelCase = {} __UpperCamelCase = torch.ones( size=(self.model_tester.n_targets,) , device=__A , dtype=torch.long ) __UpperCamelCase = torch.ones( self.model_tester.n_targets , 4 , device=__A , dtype=torch.float ) labels.append(__A ) __UpperCamelCase = labels return inputs_dict def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = YolosModelTester(self ) __UpperCamelCase = ConfigTester(self , config_class=__A , has_text_modality=__A , hidden_size=3_7 ) def _lowerCamelCase ( self : List[str] ): self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): # YOLOS does not use inputs_embeds pass def _lowerCamelCase ( self : Any ): __UpperCamelCase , __UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase = model_class(__A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__A , nn.Linear ) ) def _lowerCamelCase ( self : Any ): __UpperCamelCase , __UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase = model_class(__A ) __UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCamelCase = [*signature.parameters.keys()] __UpperCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , __A ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase , __UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase = True # in YOLOS, the seq_len is different __UpperCamelCase = self.model_tester.expected_seq_len for model_class in self.all_model_classes: __UpperCamelCase = True __UpperCamelCase = False __UpperCamelCase = True __UpperCamelCase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): __UpperCamelCase = model(**self._prepare_for_class(__A , __A ) ) __UpperCamelCase = outputs.attentions self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCamelCase = True __UpperCamelCase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): __UpperCamelCase = model(**self._prepare_for_class(__A , __A ) ) __UpperCamelCase = outputs.attentions self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) __UpperCamelCase = len(__A ) # Check attention is always last and order is fine __UpperCamelCase = True __UpperCamelCase = True __UpperCamelCase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): __UpperCamelCase = model(**self._prepare_for_class(__A , __A ) ) __UpperCamelCase = 1 self.assertEqual(out_len + added_hidden_states , len(__A ) ) __UpperCamelCase = outputs.attentions self.assertEqual(len(__A ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def _lowerCamelCase ( self : str ): def check_hidden_states_output(__A : List[str] , __A : int , __A : Tuple ): __UpperCamelCase = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): __UpperCamelCase = model(**self._prepare_for_class(__A , __A ) ) __UpperCamelCase = outputs.hidden_states __UpperCamelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__A ) , __A ) # YOLOS has a different seq_length __UpperCamelCase = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) __UpperCamelCase , __UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase = True check_hidden_states_output(__A , __A , __A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCamelCase = True check_hidden_states_output(__A , __A , __A ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*__A ) @slow def _lowerCamelCase ( self : List[str] ): for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase = YolosModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def lowercase__ ( ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class snake_case ( unittest.TestCase ): """simple docstring""" @cached_property def _lowerCamelCase ( self : Any ): return AutoImageProcessor.from_pretrained('hustvl/yolos-small' ) if is_vision_available() else None @slow def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = YolosForObjectDetection.from_pretrained('hustvl/yolos-small' ).to(__A ) __UpperCamelCase = self.default_image_processor __UpperCamelCase = prepare_img() __UpperCamelCase = image_processor(images=__A , return_tensors='pt' ).to(__A ) # forward pass with torch.no_grad(): __UpperCamelCase = model(inputs.pixel_values ) # verify outputs __UpperCamelCase = torch.Size((1, 1_0_0, 9_2) ) self.assertEqual(outputs.logits.shape , __A ) __UpperCamelCase = torch.tensor( [[-24.0248, -10.3024, -14.8290], [-42.0392, -16.8200, -27.4334], [-27.2743, -11.8154, -18.7148]] , device=__A , ) __UpperCamelCase = torch.tensor( [[0.2559, 0.5455, 0.4706], [0.2989, 0.7279, 0.1875], [0.7732, 0.4017, 0.4462]] , device=__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , __A , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , __A , atol=1e-4 ) ) # verify postprocessing __UpperCamelCase = image_processor.post_process_object_detection( __A , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] __UpperCamelCase = torch.tensor([0.9994, 0.9790, 0.9964, 0.9972, 0.9861] ).to(__A ) __UpperCamelCase = [7_5, 7_5, 1_7, 6_3, 1_7] __UpperCamelCase = torch.tensor([335.0609, 79.3848, 375.4216, 187.2495] ).to(__A ) self.assertEqual(len(results['scores'] ) , 5 ) self.assertTrue(torch.allclose(results['scores'] , __A , atol=1e-4 ) ) self.assertSequenceEqual(results['labels'].tolist() , __A ) self.assertTrue(torch.allclose(results['boxes'][0, :] , __A ) )
53
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : List[Any] ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] =("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , __A : Optional[int] , __A : List[Any] ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : List[Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : List[Any] , *__A : Dict , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : List[Any] , *__A : List[str] , **__A : Dict ): return self.tokenizer.decode(*__A , **__A ) @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
1
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig a__ : List[str] =logging.get_logger(__name__) a__ : str ={ '''Intel/dpt-large''': '''https://huggingface.co/Intel/dpt-large/resolve/main/config.json''', # See all DPT models at https://huggingface.co/models?filter=dpt } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] ="dpt" def __init__( self : int , __A : Optional[int]=7_6_8 , __A : Any=1_2 , __A : Tuple=1_2 , __A : Optional[Any]=3_0_7_2 , __A : List[Any]="gelu" , __A : Tuple=0.0 , __A : List[Any]=0.0 , __A : List[Any]=0.02 , __A : Dict=1e-12 , __A : List[Any]=3_8_4 , __A : List[str]=1_6 , __A : Union[str, Any]=3 , __A : Optional[int]=False , __A : str=True , __A : Union[str, Any]=[2, 5, 8, 1_1] , __A : Any="project" , __A : List[str]=[4, 2, 1, 0.5] , __A : Union[str, Any]=[9_6, 1_9_2, 3_8_4, 7_6_8] , __A : List[str]=2_5_6 , __A : str=-1 , __A : Dict=False , __A : Tuple=True , __A : Union[str, Any]=0.4 , __A : int=2_5_5 , __A : Tuple=0.1 , __A : Optional[Any]=[1, 1_0_2_4, 2_4, 2_4] , __A : List[str]=[0, 1] , __A : Optional[Any]=None , **__A : Tuple , ): super().__init__(**__A ) __UpperCamelCase = hidden_size __UpperCamelCase = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info('Initializing the config with a `BiT` backbone.' ) __UpperCamelCase = { 'global_padding': 'same', 'layer_type': 'bottleneck', 'depths': [3, 4, 9], 'out_features': ['stage1', 'stage2', 'stage3'], 'embedding_dynamic_padding': True, } __UpperCamelCase = BitConfig(**__A ) elif isinstance(__A , __A ): logger.info('Initializing the config with a `BiT` backbone.' ) __UpperCamelCase = BitConfig(**__A ) elif isinstance(__A , __A ): __UpperCamelCase = backbone_config else: raise ValueError( f'''backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}.''' ) __UpperCamelCase = backbone_featmap_shape __UpperCamelCase = neck_ignore_stages if readout_type != "project": raise ValueError('Readout type must be \'project\' when using `DPT-hybrid` mode.' ) else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = [] __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = initializer_range __UpperCamelCase = layer_norm_eps __UpperCamelCase = image_size __UpperCamelCase = patch_size __UpperCamelCase = num_channels __UpperCamelCase = qkv_bias __UpperCamelCase = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError('Readout_type must be one of [\'ignore\', \'add\', \'project\']' ) __UpperCamelCase = readout_type __UpperCamelCase = reassemble_factors __UpperCamelCase = neck_hidden_sizes __UpperCamelCase = fusion_hidden_size __UpperCamelCase = head_in_index __UpperCamelCase = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) __UpperCamelCase = use_auxiliary_head __UpperCamelCase = auxiliary_loss_weight __UpperCamelCase = semantic_loss_ignore_index __UpperCamelCase = semantic_classifier_dropout def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __UpperCamelCase = self.backbone_config.to_dict() __UpperCamelCase = self.__class__.model_type return output
53
'''simple docstring''' from __future__ import annotations from typing import Any class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case : """simple docstring""" def __init__( self : List[Any] , __A : Any ): __UpperCamelCase = data __UpperCamelCase = None def __iter__( self : Optional[Any] ): __UpperCamelCase = self __UpperCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(__A ) yield node.data __UpperCamelCase = node.next_node @property def _lowerCamelCase ( self : List[str] ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": a__ : Dict =Node(1) a__ : Optional[int] =Node(2) a__ : List[str] =Node(3) a__ : Optional[int] =Node(4) print(root_node.has_loop) # False a__ : str =root_node.next_node print(root_node.has_loop) # True a__ : Optional[int] =Node(5) a__ : List[Any] =Node(6) a__ : int =Node(5) a__ : Tuple =Node(6) print(root_node.has_loop) # False a__ : str =Node(1) print(root_node.has_loop) # False
53
1
'''simple docstring''' a__ : List[Any] ={0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} a__ : Optional[int] ={0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowercase__ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: """simple docstring""" __UpperCamelCase = True __UpperCamelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__lowercase , __lowercase , __lowercase ) order.append(__lowercase ) return order def lowercase__ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: """simple docstring""" __UpperCamelCase = True __UpperCamelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__lowercase , __lowercase , __lowercase ) return component def lowercase__ ( __lowercase : dict[int, list[int]] ) -> list[list[int]]: """simple docstring""" __UpperCamelCase = len(__lowercase ) * [False] __UpperCamelCase = {vert: [] for vert in range(len(__lowercase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__lowercase ) __UpperCamelCase = [] for i, was_visited in enumerate(__lowercase ): if not was_visited: order += topology_sort(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = [] __UpperCamelCase = len(__lowercase ) * [False] for i in range(len(__lowercase ) ): __UpperCamelCase = order[len(__lowercase ) - i - 1] if not visited[vert]: __UpperCamelCase = find_components(__lowercase , __lowercase , __lowercase ) components_list.append(__lowercase ) return components_list
53
'''simple docstring''' a__ : Optional[Any] =256 # Modulus to hash a string a__ : Dict =1_000_003 def lowercase__ ( __lowercase : str , __lowercase : str ) -> bool: """simple docstring""" __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) if p_len > t_len: return False __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 1 # Calculating the hash of pattern and substring of text for i in range(__lowercase ): __UpperCamelCase = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus __UpperCamelCase = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue __UpperCamelCase = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash __UpperCamelCase = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: """simple docstring""" __UpperCamelCase = 'abc1abc12' __UpperCamelCase = 'alskfjaldsabc1abc1abc12k23adsfabcabc' __UpperCamelCase = 'alskfjaldsk23adsfabcabc' assert rabin_karp(__lowercase , __lowercase ) and not rabin_karp(__lowercase , __lowercase ) # Test 2) __UpperCamelCase = 'ABABX' __UpperCamelCase = 'ABABZABABYABABX' assert rabin_karp(__lowercase , __lowercase ) # Test 3) __UpperCamelCase = 'AAAB' __UpperCamelCase = 'ABAAAAAB' assert rabin_karp(__lowercase , __lowercase ) # Test 4) __UpperCamelCase = 'abcdabcy' __UpperCamelCase = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(__lowercase , __lowercase ) # Test 5) __UpperCamelCase = 'Lü' __UpperCamelCase = 'Lüsai' assert rabin_karp(__lowercase , __lowercase ) __UpperCamelCase = 'Lue' assert not rabin_karp(__lowercase , __lowercase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
1
'''simple docstring''' import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Any , __lowercase : str , __lowercase : Dict ) -> List[Any]: """simple docstring""" if gpta_config_file == "": __UpperCamelCase = GPTaConfig() else: __UpperCamelCase = GPTaConfig.from_json_file(__lowercase ) __UpperCamelCase = GPTaModel(__lowercase ) # Load weights from numpy load_tf_weights_in_gpta(__lowercase , __lowercase , __lowercase ) # Save pytorch-model __UpperCamelCase = pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase = pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict() , __lowercase ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__lowercase , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": a__ : Optional[int] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--gpt2_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--gpt2_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained OpenAI model. \n''' '''This specifies the model architecture.''' ), ) a__ : Optional[Any] =parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
53
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor a__ : Union[str, Any] =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[Any] , *__A : str , **__A : Optional[Any] ): warnings.warn( 'The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use GLPNImageProcessor instead.' , __A , ) super().__init__(*__A , **__A )
53
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
1
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def lowercase__ ( __lowercase : Dict=None ) -> str: """simple docstring""" __UpperCamelCase = argparse.ArgumentParser(add_help=__lowercase , allow_abbrev=__lowercase ) # The main config parser __UpperCamelCase = config_command_parser(__lowercase ) # The subparser to add commands to __UpperCamelCase = config_parser.add_subparsers(title='subcommands' , dest='subcommand' ) # Then add other parsers with the parent parser default_command_parser(__lowercase , parents=[parent_parser] ) update_command_parser(__lowercase , parents=[parent_parser] ) return config_parser def lowercase__ ( ) -> int: """simple docstring""" __UpperCamelCase = get_config_parser() __UpperCamelCase = config_parser.parse_args() if not hasattr(__lowercase , 'func' ): config_parser.print_help() exit(1 ) # Run args.func(__lowercase ) if __name__ == "__main__": main()
53
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
1
'''simple docstring''' import string def lowercase__ ( __lowercase : str ) -> str: """simple docstring""" __UpperCamelCase = '' for i in sequence: __UpperCamelCase = ord(__lowercase ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def lowercase__ ( __lowercase : str ) -> str: """simple docstring""" __UpperCamelCase = string.ascii_letters __UpperCamelCase = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(__lowercase )] if c in letters else c for c in sequence ) def lowercase__ ( ) -> None: """simple docstring""" from timeit import timeit print('Running performance benchmarks...' ) __UpperCamelCase = 'from string import printable ; from __main__ import atbash, atbash_slow' print(F'''> atbash_slow(): {timeit('atbash_slow(printable)' , setup=__lowercase )} seconds''' ) print(F'''> atbash(): {timeit('atbash(printable)' , setup=__lowercase )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f'{example} encrypted in atbash: {atbash(example)}') benchmark()
53
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def lowercase__ ( __lowercase : Tuple ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = emb.weight.shape __UpperCamelCase = nn.Linear(__lowercase , __lowercase , bias=__lowercase ) __UpperCamelCase = emb.weight.data return lin_layer def lowercase__ ( __lowercase : int , __lowercase : List[str]="facebook/mbart-large-en-ro" , __lowercase : str=False , __lowercase : List[Any]=False ) -> int: """simple docstring""" __UpperCamelCase = torch.load(__lowercase , map_location='cpu' )['model'] remove_ignore_keys_(__lowercase ) __UpperCamelCase = state_dict['encoder.embed_tokens.weight'].shape[0] __UpperCamelCase = MBartConfig.from_pretrained(__lowercase , vocab_size=__lowercase ) if mbart_aa and finetuned: __UpperCamelCase = 'relu' __UpperCamelCase = state_dict['decoder.embed_tokens.weight'] __UpperCamelCase = MBartForConditionalGeneration(__lowercase ) model.model.load_state_dict(__lowercase ) if finetuned: __UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": a__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') a__ : Union[str, Any] =parser.parse_args() a__ : str =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
53
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ : int ={'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str =['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[Any] =['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Union[str, Any] =[ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[Any] =[ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a__ : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure)
53
'''simple docstring''' import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Any , __A : Dict , __A : str , __A : List[Any]=1_0_2_4 , __A : Tuple=1_0_2_4 , __A : str=3.6 ): __UpperCamelCase = tokenizer __UpperCamelCase = tokenizer.bos_token_id __UpperCamelCase = dataset __UpperCamelCase = seq_length __UpperCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self : Any ): __UpperCamelCase = iter(self.dataset ) __UpperCamelCase = True while more_examples: __UpperCamelCase , __UpperCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__A )['content'] ) buffer_len += len(buffer[-1] ) except StopIteration: __UpperCamelCase = False break __UpperCamelCase = tokenizer(__A , truncation=__A )['input_ids'] __UpperCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__A ) , self.seq_length ): __UpperCamelCase = all_token_ids[i : i + self.seq_length] if len(__A ) == self.seq_length: yield torch.tensor(__A ) def lowercase__ ( __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = {'streaming': True} __UpperCamelCase = load_dataset(args.dataset_name , split='train' , **__lowercase ) __UpperCamelCase = ConstantLengthDataset(__lowercase , __lowercase , seq_length=args.seq_length ) __UpperCamelCase = DataLoader(__lowercase , batch_size=args.batch_size ) return eval_dataloader def lowercase__ ( __lowercase : Tuple ) -> Optional[Any]: """simple docstring""" model.eval() __UpperCamelCase = [] for step, batch in enumerate(__lowercase ): with torch.no_grad(): __UpperCamelCase = model(__lowercase , labels=__lowercase ) __UpperCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(__lowercase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __UpperCamelCase = torch.mean(torch.cat(__lowercase ) ) try: __UpperCamelCase = torch.exp(__lowercase ) except OverflowError: __UpperCamelCase = float('inf' ) return loss.item(), perplexity.item() # Setup Accelerator a__ : int =Accelerator() # Parse configuration a__ : Dict =HfArgumentParser(EvaluationArguments) a__ : Union[str, Any] =parser.parse_args() set_seed(args.seed) # Logging a__ : List[Any] =logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer a__ : Union[str, Any] =AutoModelForCausalLM.from_pretrained(args.model_ckpt) a__ : List[Any] =AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a__ : Union[str, Any] =create_dataloader(args) # Prepare everything with our `accelerator`. a__ , a__ : List[str] =accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') a__ , a__ : Any =evaluate(args) logger.info(f'loss/eval: {eval_loss}, perplexity: {perplexity}')
53
1
'''simple docstring''' import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class snake_case ( __lowerCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] =GPTaTokenizer SCREAMING_SNAKE_CASE_ : int =GPTaTokenizerFast SCREAMING_SNAKE_CASE_ : Union[str, Any] =True SCREAMING_SNAKE_CASE_ : Optional[int] ={"add_prefix_space": True} SCREAMING_SNAKE_CASE_ : Optional[Any] =False def _lowerCamelCase ( self : Any ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple , **__A : Any ): kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **__A ) def _lowerCamelCase ( self : Union[str, Any] , **__A : List[str] ): kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **__A ) def _lowerCamelCase ( self : Optional[Any] , __A : str ): __UpperCamelCase = 'lower newer' __UpperCamelCase = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : Dict ): __UpperCamelCase = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __UpperCamelCase = 'lower newer' __UpperCamelCase = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] __UpperCamelCase = tokenizer.tokenize(__A , add_prefix_space=__A ) self.assertListEqual(__A , __A ) __UpperCamelCase = tokens + [tokenizer.unk_token] __UpperCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _lowerCamelCase ( self : List[Any] ): if not self.test_rust_tokenizer: return __UpperCamelCase = self.get_tokenizer() __UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=__A ) __UpperCamelCase = 'lower newer' # Testing tokenization __UpperCamelCase = tokenizer.tokenize(__A , add_prefix_space=__A ) __UpperCamelCase = rust_tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) # Testing conversion to ids without special tokens __UpperCamelCase = tokenizer.encode(__A , add_special_tokens=__A , add_prefix_space=__A ) __UpperCamelCase = rust_tokenizer.encode(__A , add_special_tokens=__A ) self.assertListEqual(__A , __A ) # Testing conversion to ids with special tokens __UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=__A ) __UpperCamelCase = tokenizer.encode(__A , add_prefix_space=__A ) __UpperCamelCase = rust_tokenizer.encode(__A ) self.assertListEqual(__A , __A ) # Testing the unknown token __UpperCamelCase = tokens + [rust_tokenizer.unk_token] __UpperCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__A ) , __A ) def _lowerCamelCase ( self : Any , *__A : Union[str, Any] , **__A : Optional[Any] ): # It's very difficult to mix/test pretokenization with byte-level # And get both GPT2 and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def _lowerCamelCase ( self : List[str] , __A : Any=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input __UpperCamelCase = 'This is a simple input' __UpperCamelCase = ['This is a simple input 1', 'This is a simple input 2'] __UpperCamelCase = ('This is a simple input', 'This is a pair') __UpperCamelCase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding='max_length' ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding='max_length' ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding='max_length' , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding='max_length' ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding='max_length' ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding='max_length' , ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input __UpperCamelCase = 'This is a simple input' __UpperCamelCase = ['This is a simple input looooooooong', 'This is a simple input'] __UpperCamelCase = ('This is a simple input', 'This is a pair') __UpperCamelCase = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] __UpperCamelCase = tokenizer.pad_token_id __UpperCamelCase = tokenizer(__A , padding='max_length' , max_length=3_0 , return_tensors='np' ) __UpperCamelCase = tokenizer(__A , padding=__A , truncate=__A , return_tensors='np' ) __UpperCamelCase = tokenizer(*__A , padding='max_length' , max_length=6_0 , return_tensors='np' ) __UpperCamelCase = tokenizer(__A , padding=__A , truncate=__A , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = '$$$' __UpperCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=__A , add_bos_token=__A ) __UpperCamelCase = 'This is a simple input' __UpperCamelCase = ['This is a simple input 1', 'This is a simple input 2'] __UpperCamelCase = tokenizer.bos_token_id __UpperCamelCase = tokenizer(__A ) __UpperCamelCase = tokenizer(__A ) self.assertEqual(out_s.input_ids[0] , __A ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __UpperCamelCase = tokenizer.decode(out_s.input_ids ) __UpperCamelCase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , __A ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def _lowerCamelCase ( self : int ): pass def _lowerCamelCase ( self : int ): # TODO: change to self.get_tokenizers() when the fast version is implemented __UpperCamelCase = [self.get_tokenizer(do_lower_case=__A , add_bos_token=__A )] for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCamelCase = 'Encode this.' __UpperCamelCase = 'This one too please.' __UpperCamelCase = tokenizer.encode(__A , add_special_tokens=__A ) encoded_sequence += tokenizer.encode(__A , add_special_tokens=__A ) __UpperCamelCase = tokenizer.encode_plus( __A , __A , add_special_tokens=__A , return_special_tokens_mask=__A , ) __UpperCamelCase = encoded_sequence_dict['input_ids'] __UpperCamelCase = encoded_sequence_dict['special_tokens_mask'] self.assertEqual(len(__A ) , len(__A ) ) __UpperCamelCase = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(__A ) ] __UpperCamelCase = [x for x in filtered_sequence if x is not None] self.assertEqual(__A , __A ) @require_tokenizers class snake_case ( unittest.TestCase ): """simple docstring""" def _lowerCamelCase ( self : Tuple ): # More context: # https://huggingface.co/wjmcat/opt-350m-paddle/discussions/1 # https://huggingface.slack.com/archives/C01N44FJDHT/p1653511495183519 # https://github.com/huggingface/transformers/pull/17088#discussion_r871246439 __UpperCamelCase = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=__A ) __UpperCamelCase = 'A photo of a cat' __UpperCamelCase = tokenizer.encode( __A , ) self.assertEqual(__A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('test_opt' ) __UpperCamelCase = AutoTokenizer.from_pretrained('./test_opt' ) __UpperCamelCase = tokenizer.encode( __A , ) self.assertEqual(__A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = AutoTokenizer.from_pretrained('facebook/opt-350m' , use_slow=__A ) __UpperCamelCase = 'A photo of a cat' __UpperCamelCase = tokenizer.encode( __A , ) # Same as above self.assertEqual(__A , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) @unittest.skip('This test is failing because of a bug in the fast tokenizer' ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=__A ) __UpperCamelCase = 'bos' __UpperCamelCase = tokenizer.get_vocab()['bos'] __UpperCamelCase = 'A photo of a cat' __UpperCamelCase = tokenizer.encode( __A , ) # We changed the bos token self.assertEqual(__A , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('./tok' ) __UpperCamelCase = AutoTokenizer.from_pretrained('./tok' ) self.assertTrue(tokenizer.is_fast ) __UpperCamelCase = tokenizer.encode( __A , ) self.assertEqual(__A , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] )
53
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging a__ : Any =logging.get_logger(__name__) a__ : Optional[Any] ={ '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict ="gpt_neo" SCREAMING_SNAKE_CASE_ : Optional[int] =["past_key_values"] SCREAMING_SNAKE_CASE_ : List[Any] ={"num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self : Union[str, Any] , __A : Union[str, Any]=5_0_2_5_7 , __A : Any=2_0_4_8 , __A : Optional[Any]=2_0_4_8 , __A : Any=2_4 , __A : Union[str, Any]=[[["global", "local"], 1_2]] , __A : str=1_6 , __A : Optional[int]=None , __A : Union[str, Any]=2_5_6 , __A : Any="gelu_new" , __A : Dict=0.0 , __A : Optional[int]=0.0 , __A : int=0.0 , __A : List[str]=0.1 , __A : Any=1e-5 , __A : int=0.02 , __A : List[str]=True , __A : Tuple=5_0_2_5_6 , __A : Optional[Any]=5_0_2_5_6 , **__A : Optional[Any] , ): __UpperCamelCase = vocab_size __UpperCamelCase = max_position_embeddings __UpperCamelCase = hidden_size __UpperCamelCase = num_layers __UpperCamelCase = num_heads __UpperCamelCase = intermediate_size __UpperCamelCase = window_size __UpperCamelCase = activation_function __UpperCamelCase = resid_dropout __UpperCamelCase = embed_dropout __UpperCamelCase = attention_dropout __UpperCamelCase = classifier_dropout __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = initializer_range __UpperCamelCase = use_cache __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id __UpperCamelCase = attention_types __UpperCamelCase = self.expand_attention_types_params(__A ) if len(self.attention_layers ) != self.num_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.attention_layers)` == `config.num_layers` ' f'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' f'''`config.num_layers = {self.num_layers}`. ''' '`config.attention_layers` is prepared using `config.attention_types`. ' 'Please verify the value of `config.attention_types` argument.' ) super().__init__(bos_token_id=__A , eos_token_id=__A , **__A ) @staticmethod def _lowerCamelCase ( __A : Tuple ): __UpperCamelCase = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase__ ( __lowercase : Tuple , __lowercase : Any , __lowercase : Union[str, Any] , __lowercase : List[str] ) -> Any: """simple docstring""" import torch __UpperCamelCase = input.size() __UpperCamelCase = len(__lowercase ) __UpperCamelCase = shape[dimension] __UpperCamelCase = torch.arange(0 , __lowercase , __lowercase ) __UpperCamelCase = torch.div(sizedim - size , __lowercase , rounding_mode='floor' ) + 1 __UpperCamelCase = torch.arange(__lowercase ) + low_indices[:min_length][:, None] __UpperCamelCase = [slice(__lowercase )] * rank __UpperCamelCase = indices __UpperCamelCase = input[s] __UpperCamelCase = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" import torch __UpperCamelCase = torch.arange(1 , __lowercase ) __UpperCamelCase = torch.remainder(__lowercase , __lowercase ) __UpperCamelCase = remainders == 0 __UpperCamelCase = candidates[divisor_indices] __UpperCamelCase = torch.max(__lowercase ) return largest_divisor, torch.div(__lowercase , __lowercase , rounding_mode='floor' ) class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(__A , direction='inputs' ) __UpperCamelCase = {0: 'batch', 1: 'past_sequence + sequence'} else: __UpperCamelCase = {0: 'batch', 1: 'sequence'} return common_inputs @property def _lowerCamelCase ( self : int ): return self._config.num_heads def _lowerCamelCase ( self : List[str] , __A : PreTrainedTokenizer , __A : int = -1 , __A : int = -1 , __A : bool = False , __A : Optional[TensorType] = None , ): __UpperCamelCase = super(__A , self ).generate_dummy_inputs( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) # We need to order the input in the way they appears in the forward() __UpperCamelCase = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __UpperCamelCase , __UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values __UpperCamelCase = seqlen + 2 __UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __UpperCamelCase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(self.num_layers ) ] __UpperCamelCase = common_inputs['attention_mask'] if self.use_past: __UpperCamelCase = ordered_inputs['attention_mask'].dtype __UpperCamelCase = torch.cat( [ordered_inputs['attention_mask'], torch.ones(__A , __A , dtype=__A )] , dim=1 ) return ordered_inputs @property def _lowerCamelCase ( self : Dict ): return 1_3
53
1
'''simple docstring''' def lowercase__ ( __lowercase : int , __lowercase : int ) -> int: """simple docstring""" while a != 0: __UpperCamelCase , __UpperCamelCase = b % a, a return b def lowercase__ ( __lowercase : int , __lowercase : int ) -> int: """simple docstring""" if gcd(__lowercase , __lowercase ) != 1: __UpperCamelCase = F'''mod inverse of {a!r} and {m!r} does not exist''' raise ValueError(__lowercase ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1, 0, a __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 0, 1, m while va != 0: __UpperCamelCase = ua // va __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
53
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="naver-clova-ix/donut-base-finetuned-docvqa" SCREAMING_SNAKE_CASE_ : Dict =( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) SCREAMING_SNAKE_CASE_ : List[str] ="document_qa" SCREAMING_SNAKE_CASE_ : Union[str, Any] =AutoProcessor SCREAMING_SNAKE_CASE_ : Union[str, Any] =VisionEncoderDecoderModel SCREAMING_SNAKE_CASE_ : List[Any] =["image", "text"] SCREAMING_SNAKE_CASE_ : Any =["text"] def __init__( self : Optional[int] , *__A : List[str] , **__A : List[Any] ): if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.' ) super().__init__(*__A , **__A ) def _lowerCamelCase ( self : Any , __A : "Image" , __A : str ): __UpperCamelCase = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' __UpperCamelCase = task_prompt.replace('{user_input}' , __A ) __UpperCamelCase = self.pre_processor.tokenizer( __A , add_special_tokens=__A , return_tensors='pt' ).input_ids __UpperCamelCase = self.pre_processor(__A , return_tensors='pt' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _lowerCamelCase ( self : Union[str, Any] , __A : Optional[Any] ): return self.model.generate( inputs['pixel_values'].to(self.device ) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__A , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__A , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__A , ).sequences def _lowerCamelCase ( self : Tuple , __A : List[Any] ): __UpperCamelCase = self.pre_processor.batch_decode(__A )[0] __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '' ) __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '' ) __UpperCamelCase = re.sub(R'<.*?>' , '' , __A , count=1 ).strip() # remove first task start token __UpperCamelCase = self.pre_processor.tokenajson(__A ) return sequence["answer"]
53
1
'''simple docstring''' print((lambda quine: quine % quine)('''print((lambda quine: quine %% quine)(%r))'''))
53
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
53
1
'''simple docstring''' def lowercase__ ( __lowercase : list[int] ) -> int: """simple docstring""" if not numbers: return 0 if not isinstance(__lowercase , (list, tuple) ) or not all( isinstance(__lowercase , __lowercase ) for number in numbers ): raise ValueError('numbers must be an iterable of integers' ) __UpperCamelCase = __UpperCamelCase = __UpperCamelCase = numbers[0] for i in range(1 , len(__lowercase ) ): # update the maximum and minimum subarray products __UpperCamelCase = numbers[i] if number < 0: __UpperCamelCase , __UpperCamelCase = min_till_now, max_till_now __UpperCamelCase = max(__lowercase , max_till_now * number ) __UpperCamelCase = min(__lowercase , min_till_now * number ) # update the maximum product found till now __UpperCamelCase = max(__lowercase , __lowercase ) return max_prod
53
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowercase__ ( __lowercase : Features ) -> Optional[int]: """simple docstring""" __UpperCamelCase = np.inf def set_batch_size(__lowercase : FeatureType ) -> None: nonlocal batch_size if isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__lowercase , __lowercase ) and feature.dtype == "binary": __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__lowercase , __lowercase ) return None if batch_size is np.inf else batch_size class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , __A : NestedDataStructureLike[PathLike] , __A : Optional[NamedSplit] = None , __A : Optional[Features] = None , __A : str = None , __A : bool = False , __A : bool = False , __A : Optional[int] = None , **__A : Dict , ): super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) __UpperCamelCase = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} __UpperCamelCase = _PACKAGED_DATASETS_MODULES['parquet'][1] __UpperCamelCase = Parquet( cache_dir=__A , data_files=__A , features=__A , hash=__A , **__A , ) def _lowerCamelCase ( self : Optional[int] ): # Build iterable dataset if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Dataset , __A : Union[PathLike, BinaryIO] , __A : Optional[int] = None , **__A : Dict , ): __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size or get_writer_batch_size(dataset.features ) __UpperCamelCase = parquet_writer_kwargs def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , 'wb+' ) as buffer: __UpperCamelCase = self._write(file_obj=__A , batch_size=__A , **self.parquet_writer_kwargs ) else: __UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__A , **self.parquet_writer_kwargs ) return written def _lowerCamelCase ( self : List[str] , __A : BinaryIO , __A : int , **__A : List[str] ): __UpperCamelCase = 0 __UpperCamelCase = parquet_writer_kwargs.pop('path_or_buf' , __A ) __UpperCamelCase = self.dataset.features.arrow_schema __UpperCamelCase = pq.ParquetWriter(__A , schema=__A , **__A ) for offset in logging.tqdm( range(0 , len(self.dataset ) , __A ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating parquet from Arrow format' , ): __UpperCamelCase = query_table( table=self.dataset._data , key=slice(__A , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__A ) written += batch.nbytes writer.close() return written
53
1
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def lowercase__ ( __lowercase : Optional[int]=None ) -> Union[str, Any]: """simple docstring""" if subparsers is not None: __UpperCamelCase = subparsers.add_parser('env' ) else: __UpperCamelCase = argparse.ArgumentParser('Accelerate env command' ) parser.add_argument( '--config_file' , default=__lowercase , help='The config file to use for the default values in the launching script.' ) if subparsers is not None: parser.set_defaults(func=__lowercase ) return parser def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" __UpperCamelCase = torch.__version__ __UpperCamelCase = torch.cuda.is_available() __UpperCamelCase = is_xpu_available() __UpperCamelCase = is_npu_available() __UpperCamelCase = 'Not found' # Get the default from the config file. if args.config_file is not None or os.path.isfile(__lowercase ): __UpperCamelCase = load_config_from_file(args.config_file ).to_dict() __UpperCamelCase = { '`Accelerate` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'Numpy version': np.__version__, 'PyTorch version (GPU?)': F'''{pt_version} ({pt_cuda_available})''', 'PyTorch XPU available': str(__lowercase ), 'PyTorch NPU available': str(__lowercase ), 'System RAM': F'''{psutil.virtual_memory().total / 1024 ** 3:.2f} GB''', } if pt_cuda_available: __UpperCamelCase = torch.cuda.get_device_name() print('\nCopy-and-paste the text below in your GitHub issue\n' ) print('\n'.join([F'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('- `Accelerate` default config:' if args.config_file is None else '- `Accelerate` config passed:' ) __UpperCamelCase = ( '\n'.join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(__lowercase , __lowercase ) else F'''\t{accelerate_config}''' ) print(__lowercase ) __UpperCamelCase = accelerate_config return info def lowercase__ ( ) -> int: """simple docstring""" __UpperCamelCase = env_command_parser() __UpperCamelCase = parser.parse_args() env_command(__lowercase ) return 0 if __name__ == "__main__": raise SystemExit(main())
53
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def lowercase__ ( __lowercase : SplitDict ) -> int: """simple docstring""" __UpperCamelCase = split_dict._to_yaml_list() assert len(__lowercase ) == len(__lowercase ) __UpperCamelCase = SplitDict._from_yaml_list(__lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __UpperCamelCase = None # the split name of split_dict takes over the name of the split info object __UpperCamelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=__lowercase ), SplitInfo(dataset_name='my_dataset' )] ) def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" __UpperCamelCase = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
53
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Optional[int] =logging.get_logger(__name__) a__ : Optional[int] ={ '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="mgp-str" def __init__( self : List[str] , __A : Tuple=[3_2, 1_2_8] , __A : Optional[int]=4 , __A : str=3 , __A : Any=2_7 , __A : Optional[Any]=3_8 , __A : int=5_0_2_5_7 , __A : int=3_0_5_2_2 , __A : Tuple=7_6_8 , __A : str=1_2 , __A : List[Any]=1_2 , __A : List[Any]=4.0 , __A : Dict=True , __A : List[str]=False , __A : Optional[Any]=1e-5 , __A : Any=0.0 , __A : int=0.0 , __A : Optional[Any]=0.0 , __A : str=False , __A : List[Any]=0.02 , **__A : List[Any] , ): super().__init__(**__A ) __UpperCamelCase = image_size __UpperCamelCase = patch_size __UpperCamelCase = num_channels __UpperCamelCase = max_token_length __UpperCamelCase = num_character_labels __UpperCamelCase = num_bpe_labels __UpperCamelCase = num_wordpiece_labels __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = mlp_ratio __UpperCamelCase = distilled __UpperCamelCase = layer_norm_eps __UpperCamelCase = drop_rate __UpperCamelCase = qkv_bias __UpperCamelCase = attn_drop_rate __UpperCamelCase = drop_path_rate __UpperCamelCase = output_aa_attentions __UpperCamelCase = initializer_range
53
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[str] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Any =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys a__ : str =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
1
'''simple docstring''' # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version a__ : List[str] =get_logger(__name__) class snake_case : """simple docstring""" SCREAMING_SNAKE_CASE_ : str ="dummy_data" SCREAMING_SNAKE_CASE_ : List[str] ="datasets" SCREAMING_SNAKE_CASE_ : Dict =False def __init__( self : str , __A : str , __A : str , __A : Union[Version, str] , __A : Optional[str] = None , __A : bool = False , __A : bool = True , __A : Optional[List[Callable]] = None , ): __UpperCamelCase = 0 __UpperCamelCase = dataset_name __UpperCamelCase = cache_dir __UpperCamelCase = use_local_dummy_data __UpperCamelCase = config # download_callbacks take a single url as input __UpperCamelCase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root __UpperCamelCase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general __UpperCamelCase = str(__A ) # to be downloaded __UpperCamelCase = None __UpperCamelCase = None @property def _lowerCamelCase ( self : int ): if self._dummy_file is None: __UpperCamelCase = self.download_dummy_data() return self._dummy_file @property def _lowerCamelCase ( self : Optional[int] ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def _lowerCamelCase ( self : Union[str, Any] ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) __UpperCamelCase = cached_path( __A , cache_dir=self.cache_dir , extract_compressed_file=__A , force_extract=__A ) return os.path.join(__A , self.dummy_file_name ) @property def _lowerCamelCase ( self : List[str] ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def _lowerCamelCase ( self : Tuple ): if self._bucket_url is None: __UpperCamelCase = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def _lowerCamelCase ( self : Union[str, Any] ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def _lowerCamelCase ( self : Optional[int] , __A : Optional[int] , *__A : Optional[int] ): if self.load_existing_dummy_data: # dummy data is downloaded and tested __UpperCamelCase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned __UpperCamelCase = self.dummy_file_name # special case when data_url is a dict if isinstance(__A , __A ): return self.create_dummy_data_dict(__A , __A ) elif isinstance(__A , (list, tuple) ): return self.create_dummy_data_list(__A , __A ) else: return self.create_dummy_data_single(__A , __A ) def _lowerCamelCase ( self : Any , __A : Dict , *__A : Any ): return self.download_and_extract(__A ) def _lowerCamelCase ( self : Optional[Any] , __A : Tuple , __A : Dict ): return self.download_and_extract(__A ) def _lowerCamelCase ( self : int , __A : Any , *__A : List[str] , **__A : Optional[Any] ): return path def _lowerCamelCase ( self : str ): return {} def _lowerCamelCase ( self : Any , __A : int , __A : Any ): __UpperCamelCase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(__A , __A ): for single_url in single_urls: download_callback(__A ) else: __UpperCamelCase = single_urls download_callback(__A ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(__A , __A ): __UpperCamelCase = [os.path.join(__A , urllib.parse.quote_plus(Path(__A ).name ) ) for x in single_urls] else: __UpperCamelCase = single_urls __UpperCamelCase = os.path.join(__A , urllib.parse.quote_plus(Path(__A ).name ) ) __UpperCamelCase = value # make sure that values are unique if all(isinstance(__A , __A ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique __UpperCamelCase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def _lowerCamelCase ( self : int , __A : Dict , __A : Optional[Any] ): __UpperCamelCase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one __UpperCamelCase = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , __A ) ) for url in data_url ) __UpperCamelCase = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): __UpperCamelCase = [data_url[0]] * len(__A ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(__A ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __UpperCamelCase = os.path.join(__A , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(__A ) return dummy_data_list def _lowerCamelCase ( self : Optional[int] , __A : Any , __A : List[str] ): for download_callback in self.download_callbacks: download_callback(__A ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __UpperCamelCase = os.path.join(__A , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(__A ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def _lowerCamelCase ( self : str ): pass def _lowerCamelCase ( self : List[Any] ): pass def _lowerCamelCase ( self : Tuple , __A : Dict ): def _iter_archive_members(__A : Tuple ): # this preserves the order of the members inside the ZIP archive __UpperCamelCase = Path(self.dummy_file ).parent __UpperCamelCase = path.relative_to(__A ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: __UpperCamelCase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(__A ) __UpperCamelCase = Path(__A ) __UpperCamelCase = _iter_archive_members(__A ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(__A ).as_posix(), file_path.open('rb' ) def _lowerCamelCase ( self : str , __A : Optional[Any] ): if not isinstance(__A , __A ): __UpperCamelCase = [paths] for path in paths: if os.path.isfile(__A ): if os.path.basename(__A ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(__A ): if os.path.basename(__A ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(__A ): if filename.startswith(('.', '__') ): continue yield os.path.join(__A , __A )
53
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a__ : str =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =["input_features", "attention_mask"] def __init__( self : Union[str, Any] , __A : Optional[int]=8_0 , __A : Tuple=1_6_0_0_0 , __A : Optional[Any]=8_0 , __A : Any=0.0 , __A : Any=True , __A : List[str]=True , __A : str=True , **__A : List[Any] , ): super().__init__(feature_size=__A , sampling_rate=__A , padding_value=__A , **__A ) __UpperCamelCase = num_mel_bins __UpperCamelCase = do_ceptral_normalize __UpperCamelCase = normalize_means __UpperCamelCase = normalize_vars __UpperCamelCase = True def _lowerCamelCase ( self : Union[str, Any] , __A : np.ndarray , ): __UpperCamelCase = waveform * (2**1_5) # Kaldi compliance: 16-bit signed integers __UpperCamelCase = torch.from_numpy(__A ).unsqueeze(0 ) __UpperCamelCase = ta_kaldi.fbank(__A , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( __A : np.ndarray , __A : int , __A : Optional[bool] = True , __A : Optional[bool] = True , __A : float = 0.0 , ): # make sure we normalize float32 arrays if normalize_means: __UpperCamelCase = x[:input_length].mean(axis=0 ) __UpperCamelCase = np.subtract(__A , __A ) if normalize_vars: __UpperCamelCase = x[:input_length].std(axis=0 ) __UpperCamelCase = np.divide(__A , __A ) if input_length < x.shape[0]: __UpperCamelCase = padding_value # make sure array is in float32 __UpperCamelCase = x.astype(np.floataa ) return x def _lowerCamelCase ( self : int , __A : List[np.ndarray] , __A : Optional[np.ndarray] = None ): __UpperCamelCase = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(__A , __A , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(__A , __A ) ] def __call__( self : List[Any] , __A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __A : Union[bool, str, PaddingStrategy] = False , __A : Optional[int] = None , __A : bool = False , __A : Optional[int] = None , __A : Optional[Union[str, TensorType]] = None , __A : Optional[int] = None , __A : Optional[bool] = None , **__A : Dict , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) __UpperCamelCase = isinstance(__A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __UpperCamelCase = is_batched_numpy or ( isinstance(__A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(__A , np.ndarray ): __UpperCamelCase = np.asarray(__A , dtype=np.floataa ) elif isinstance(__A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCamelCase = [raw_speech] # extract fbank features __UpperCamelCase = [self._extract_fbank_features(__A ) for waveform in raw_speech] # convert into correct format for padding __UpperCamelCase = BatchFeature({'input_features': features} ) __UpperCamelCase = self.pad( __A , padding=__A , max_length=__A , truncation=__A , pad_to_multiple_of=__A , return_attention_mask=__A , **__A , ) # make sure list is in array format __UpperCamelCase = padded_inputs.get('input_features' ) if isinstance(input_features[0] , __A ): __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for feature in input_features] __UpperCamelCase = padded_inputs.get('attention_mask' ) if attention_mask is not None: __UpperCamelCase = [np.asarray(__A , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __UpperCamelCase = ( np.array(__A , dtype=np.intaa ) if self._get_padding_strategies(__A , max_length=__A ) is not PaddingStrategy.DO_NOT_PAD else None ) __UpperCamelCase = self.normalize( padded_inputs['input_features'] , attention_mask=__A ) if return_tensors is not None: __UpperCamelCase = padded_inputs.convert_to_tensors(__A ) return padded_inputs
53
1
'''simple docstring''' import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version('''>=''', FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType a__ : Optional[Any] =get_logger(__name__) def lowercase__ ( __lowercase : Optional[int] , __lowercase : List[str] , __lowercase : str , __lowercase : List[Any] , __lowercase : List[str]=0 ) -> Any: """simple docstring""" os.makedirs(__lowercase , exist_ok=__lowercase ) with FSDP.state_dict_type( __lowercase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __UpperCamelCase = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __UpperCamelCase = F'''{MODEL_NAME}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}.bin''' __UpperCamelCase = os.path.join(__lowercase , __lowercase ) if accelerator.process_index == 0: logger.info(F'''Saving model to {output_model_file}''' ) torch.save(__lowercase , __lowercase ) logger.info(F'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __UpperCamelCase = ( F'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) __UpperCamelCase = os.path.join(__lowercase , __lowercase ) logger.info(F'''Saving model to {output_model_file}''' ) torch.save(__lowercase , __lowercase ) logger.info(F'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __UpperCamelCase = os.path.join(__lowercase , F'''{MODEL_NAME}_{model_index}''' ) os.makedirs(__lowercase , exist_ok=__lowercase ) logger.info(F'''Saving model to {ckpt_dir}''' ) __UpperCamelCase = {'model': state_dict} dist_cp.save_state_dict( state_dict=__lowercase , storage_writer=dist_cp.FileSystemWriter(__lowercase ) , planner=DefaultSavePlanner() , ) logger.info(F'''Model saved to {ckpt_dir}''' ) def lowercase__ ( __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : Any , __lowercase : str , __lowercase : List[str]=0 ) -> str: """simple docstring""" accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowercase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(__lowercase ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( 'Set the `sync_module_states` flag to `True` so that model states are synced across processes when ' 'initializing FSDP object' ) return __UpperCamelCase = F'''{MODEL_NAME}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}.bin''' __UpperCamelCase = os.path.join(__lowercase , __lowercase ) logger.info(F'''Loading model from {input_model_file}''' ) __UpperCamelCase = torch.load(__lowercase ) logger.info(F'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __UpperCamelCase = ( F'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) __UpperCamelCase = os.path.join(__lowercase , __lowercase ) logger.info(F'''Loading model from {input_model_file}''' ) __UpperCamelCase = torch.load(__lowercase ) logger.info(F'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __UpperCamelCase = ( os.path.join(__lowercase , F'''{MODEL_NAME}_{model_index}''' ) if F'''{MODEL_NAME}''' not in input_dir else input_dir ) logger.info(F'''Loading model from {ckpt_dir}''' ) __UpperCamelCase = {'model': model.state_dict()} dist_cp.load_state_dict( state_dict=__lowercase , storage_reader=dist_cp.FileSystemReader(__lowercase ) , planner=DefaultLoadPlanner() , ) __UpperCamelCase = state_dict['model'] logger.info(F'''Model loaded from {ckpt_dir}''' ) model.load_state_dict(__lowercase ) def lowercase__ ( __lowercase : Optional[Any] , __lowercase : Tuple , __lowercase : Optional[int] , __lowercase : Optional[Any] , __lowercase : List[Any] , __lowercase : Tuple=0 ) -> Optional[int]: """simple docstring""" os.makedirs(__lowercase , exist_ok=__lowercase ) with FSDP.state_dict_type( __lowercase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __UpperCamelCase = FSDP.optim_state_dict(__lowercase , __lowercase ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: __UpperCamelCase = ( F'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else F'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) __UpperCamelCase = os.path.join(__lowercase , __lowercase ) logger.info(F'''Saving Optimizer state to {output_optimizer_file}''' ) torch.save(__lowercase , __lowercase ) logger.info(F'''Optimizer state saved in {output_optimizer_file}''' ) else: __UpperCamelCase = os.path.join(__lowercase , F'''{OPTIMIZER_NAME}_{optimizer_index}''' ) os.makedirs(__lowercase , exist_ok=__lowercase ) logger.info(F'''Saving Optimizer state to {ckpt_dir}''' ) dist_cp.save_state_dict( state_dict={'optimizer': optim_state} , storage_writer=dist_cp.FileSystemWriter(__lowercase ) , planner=DefaultSavePlanner() , ) logger.info(F'''Optimizer state saved in {ckpt_dir}''' ) def lowercase__ ( __lowercase : Tuple , __lowercase : List[str] , __lowercase : List[Any] , __lowercase : Optional[int] , __lowercase : str , __lowercase : Union[str, Any]=0 ) -> List[Any]: """simple docstring""" accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowercase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __UpperCamelCase = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: __UpperCamelCase = ( F'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else F'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) __UpperCamelCase = os.path.join(__lowercase , __lowercase ) logger.info(F'''Loading Optimizer state from {input_optimizer_file}''' ) __UpperCamelCase = torch.load(__lowercase ) logger.info(F'''Optimizer state loaded from {input_optimizer_file}''' ) else: __UpperCamelCase = ( os.path.join(__lowercase , F'''{OPTIMIZER_NAME}_{optimizer_index}''' ) if F'''{OPTIMIZER_NAME}''' not in input_dir else input_dir ) logger.info(F'''Loading Optimizer from {ckpt_dir}''' ) __UpperCamelCase = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key='optimizer' , storage_reader=dist_cp.FileSystemReader(__lowercase ) , ) __UpperCamelCase = optim_state['optimizer'] logger.info(F'''Optimizer loaded from {ckpt_dir}''' ) __UpperCamelCase = FSDP.optim_state_dict_to_load(__lowercase , __lowercase , __lowercase ) optimizer.load_state_dict(__lowercase )
53
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : List[Any] =logging.get_logger(__name__) a__ : List[Any] ={ '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_text_model" def __init__( self : str , __A : List[Any]=2_5_0_0_0_2 , __A : Any=1_0_2_4 , __A : int=2_4 , __A : Dict=1_6 , __A : Optional[Any]=4_0_9_6 , __A : Union[str, Any]="gelu" , __A : Dict=0.1 , __A : Dict=0.1 , __A : List[str]=5_1_4 , __A : Optional[int]=1 , __A : int=0.02 , __A : Optional[Any]=0.02 , __A : Optional[Any]=1e-05 , __A : Dict=1 , __A : List[Any]=0 , __A : int=2 , __A : Tuple="absolute" , __A : Optional[Any]=True , __A : Optional[int]=7_6_8 , **__A : List[str] , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = use_cache __UpperCamelCase = project_dim class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_vision_model" def __init__( self : List[Any] , __A : Union[str, Any]=7_6_8 , __A : Optional[int]=3_0_7_2 , __A : Optional[Any]=5_1_2 , __A : Tuple=1_2 , __A : Union[str, Any]=1_2 , __A : Optional[int]=3 , __A : Dict=2_2_4 , __A : Tuple=3_2 , __A : str="quick_gelu" , __A : Dict=1e-5 , __A : Optional[int]=0.0 , __A : List[Any]=0.02 , __A : int=1.0 , **__A : Optional[int] , ): super().__init__(**__A ) __UpperCamelCase = hidden_size __UpperCamelCase = intermediate_size __UpperCamelCase = projection_dim __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = num_channels __UpperCamelCase = patch_size __UpperCamelCase = image_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = attention_dropout __UpperCamelCase = layer_norm_eps __UpperCamelCase = hidden_act @classmethod def _lowerCamelCase ( cls : Optional[Any] , __A : Union[str, os.PathLike] , **__A : Optional[Any] ): cls._set_token_in_kwargs(__A ) __UpperCamelCase , __UpperCamelCase = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('model_type' ) == "altclip": __UpperCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] ="altclip" SCREAMING_SNAKE_CASE_ : Optional[int] =True def __init__( self : Any , __A : List[str]=None , __A : List[Any]=None , __A : List[str]=7_6_8 , __A : List[str]=2.6592 , **__A : Dict ): # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __UpperCamelCase = kwargs.pop('text_config_dict' , __A ) __UpperCamelCase = kwargs.pop('vision_config_dict' , __A ) super().__init__(**__A ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __UpperCamelCase = {} # This is the complete result when using `text_config_dict`. __UpperCamelCase = AltCLIPTextConfig(**__A ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `text_config_dict` and `text_config` but with different values. ''' f'''The value `text_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The ''' f'''value `text_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __UpperCamelCase = {} # This is the complete result when using `vision_config_dict`. __UpperCamelCase = AltCLIPVisionConfig(**__A ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __UpperCamelCase = { str(__A ): value for key, value in _vision_config_dict['id2label'].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `vision_config_dict` and `vision_config` but with different ''' f'''values. The value `vision_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. ''' f'''The value `vision_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __UpperCamelCase = {} logger.info('`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.' ) if vision_config is None: __UpperCamelCase = {} logger.info('`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.' ) __UpperCamelCase = AltCLIPTextConfig(**__A ) __UpperCamelCase = AltCLIPVisionConfig(**__A ) __UpperCamelCase = projection_dim __UpperCamelCase = logit_scale_init_value __UpperCamelCase = 1.0 @classmethod def _lowerCamelCase ( cls : Union[str, Any] , __A : AltCLIPTextConfig , __A : AltCLIPVisionConfig , **__A : Optional[Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) __UpperCamelCase = self.text_config.to_dict() __UpperCamelCase = self.vision_config.to_dict() __UpperCamelCase = self.__class__.model_type return output
53
1
'''simple docstring''' from __future__ import annotations a__ : Union[str, Any] =list[list[int]] # assigning initial values to the grid a__ : Matrix =[ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution a__ : Matrix =[ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowercase__ ( __lowercase : Matrix , __lowercase : int , __lowercase : int , __lowercase : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowercase__ ( __lowercase : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowercase__ ( __lowercase : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(__lowercase ): __UpperCamelCase , __UpperCamelCase = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(__lowercase , __lowercase , __lowercase , __lowercase ): __UpperCamelCase = digit if sudoku(__lowercase ) is not None: return grid __UpperCamelCase = 0 return None def lowercase__ ( __lowercase : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(__lowercase , end=' ' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('''\nExample grid:\n''' + '''=''' * 20) print_solution(example_grid) print('''\nExample grid solution:''') a__ : List[str] =sudoku(example_grid) if solution is not None: print_solution(solution) else: print('''Cannot find a solution.''')
53
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Any ) -> Optional[Any]: """simple docstring""" with open(__lowercase ) as metadata_file: __UpperCamelCase = json.load(__lowercase ) __UpperCamelCase = LukeConfig(use_entity_aware_attention=__lowercase , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __UpperCamelCase = torch.load(__lowercase , map_location='cpu' ) # Load the entity vocab file __UpperCamelCase = load_entity_vocab(__lowercase ) __UpperCamelCase = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __UpperCamelCase = AddedToken('<ent>' , lstrip=__lowercase , rstrip=__lowercase ) __UpperCamelCase = AddedToken('<ent2>' , lstrip=__lowercase , rstrip=__lowercase ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(__lowercase , __lowercase ) __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase ) # Initialize the embeddings of the special tokens __UpperCamelCase = state_dict['embeddings.word_embeddings.weight'] __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __UpperCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __UpperCamelCase = F'''encoder.layer.{layer_index}.attention.self.''' __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __UpperCamelCase = state_dict['entity_embeddings.entity_embeddings.weight'] __UpperCamelCase = entity_emb[entity_vocab['[MASK]']] __UpperCamelCase = LukeModel(config=__lowercase ).eval() __UpperCamelCase , __UpperCamelCase = model.load_state_dict(__lowercase , strict=__lowercase ) if not (len(__lowercase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {', '.join(__lowercase )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' F''' {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}''' ) # Check outputs __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase , task='entity_classification' ) __UpperCamelCase = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __UpperCamelCase = (39, 42) __UpperCamelCase = tokenizer(__lowercase , entity_spans=[span] , add_prefix_space=__lowercase , return_tensors='pt' ) __UpperCamelCase = model(**__lowercase ) # Verify word hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 42, 1024) ) __UpperCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base __UpperCamelCase = torch.Size((1, 42, 768) ) __UpperCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 1, 1024) ) __UpperCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base __UpperCamelCase = torch.Size((1, 1, 768) ) __UpperCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(__lowercase ) ) model.save_pretrained(__lowercase ) def lowercase__ ( __lowercase : Dict ) -> List[str]: """simple docstring""" __UpperCamelCase = {} with open(__lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(__lowercase ): __UpperCamelCase , __UpperCamelCase = line.rstrip().split('\t' ) __UpperCamelCase = index return entity_vocab if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : str =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
53
1
'''simple docstring''' import argparse import importlib from pathlib import Path # Test all the extensions added in the setup a__ : Tuple =[ '''kernels/rwkv/wkv_cuda.cu''', '''kernels/rwkv/wkv_op.cpp''', '''kernels/deformable_detr/ms_deform_attn.h''', '''kernels/deformable_detr/cuda/ms_deform_im2col_cuda.cuh''', '''models/graphormer/algos_graphormer.pyx''', ] def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" for file in FILES_TO_FIND: if not (transformers_path / file).exists(): return False return True if __name__ == "__main__": a__ : Tuple =argparse.ArgumentParser() parser.add_argument('''--check_lib''', action='''store_true''', help='''Whether to check the build or the actual package.''') a__ : Union[str, Any] =parser.parse_args() if args.check_lib: a__ : Optional[int] =importlib.import_module('''transformers''') a__ : List[str] =Path(transformers_module.__file__).parent else: a__ : Tuple =Path.cwd() / '''build/lib/transformers''' if not test_custom_files_are_present(transformers_path): raise ValueError('''The built release does not contain the custom files. Fix this before going further!''')
53
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case ( __lowerCamelCase ): """simple docstring""" def _lowerCamelCase ( self : Any ): __UpperCamelCase = tempfile.mkdtemp() __UpperCamelCase = 8 # DPR tok __UpperCamelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Optional[int] ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Union[str, Any] ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def _lowerCamelCase ( self : str ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _lowerCamelCase ( self : Any , __A : bool ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __UpperCamelCase = os.path.join(self.tmpdirname , 'dataset' ) __UpperCamelCase = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _lowerCamelCase ( self : int ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __UpperCamelCase = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __UpperCamelCase = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __UpperCamelCase = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__A , open(__A , 'wb' ) ) __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_legacy_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __A ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Optional[Any] ): import torch __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) __UpperCamelCase = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors='pt' , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dpr_ctx_encoder_tokenizer() __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __A ) # check for doc token related keys in dictionary.
53
1
'''simple docstring''' import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py a__ : Union[str, Any] ='''src/transformers''' a__ : str ='''docs/source/en/tasks''' def lowercase__ ( __lowercase : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : str ) -> Tuple: """simple docstring""" with open(__lowercase , 'r' , encoding='utf-8' , newline='\n' ) as f: __UpperCamelCase = f.readlines() # Find the start prompt. __UpperCamelCase = 0 while not lines[start_index].startswith(__lowercase ): start_index += 1 start_index += 1 __UpperCamelCase = start_index while not lines[end_index].startswith(__lowercase ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. a__ : Dict =direct_transformers_import(TRANSFORMERS_PATH) a__ : Optional[Any] ={ '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). a__ : Dict ={ '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = TASK_GUIDE_TO_MODELS[task_guide] __UpperCamelCase = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(__lowercase , set() ) __UpperCamelCase = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F'''[{name}](../model_doc/{code})''' for code, name in model_names.items()] ) + "\n" def lowercase__ ( __lowercase : Any , __lowercase : Tuple=False ) -> List[Any]: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _find_text_in_file( filename=os.path.join(__lowercase , __lowercase ) , start_prompt='<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->' , end_prompt='<!--End of the generated tip-->' , ) __UpperCamelCase = get_model_list_for_task(__lowercase ) if current_list != new_list: if overwrite: with open(os.path.join(__lowercase , __lowercase ) , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F'''The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`''' ' to fix this.' ) if __name__ == "__main__": a__ : str =argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') a__ : Optional[int] =parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
53
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[Any] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys a__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
1
'''simple docstring''' def lowercase__ ( __lowercase : int ) -> list[int]: """simple docstring""" if num <= 0: raise ValueError('Input must be a positive integer' ) __UpperCamelCase = [True] * (num + 1) __UpperCamelCase = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , __lowercase ): __UpperCamelCase = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() a__ : str =int(input('''Enter a positive integer: ''').strip()) print(prime_sieve_eratosthenes(user_num))
53
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
1
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : Dict ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] ="AutoTokenizer" def __init__( self : int , __A : List[str] , __A : Tuple ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : Union[str, Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : Optional[int] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : Optional[Any] , *__A : List[str] , **__A : List[str] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : Union[str, Any] , *__A : Dict , **__A : Optional[Any] ): return self.tokenizer.decode(*__A , **__A ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
'''simple docstring''' # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys a__ : Tuple ='''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
53
1
'''simple docstring''' import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation a__ : int =logging.get_logger(__name__) a__ : List[Any] ={'''vocab_file''': '''vocab.txt''', '''emoji_file''': '''emoji.json'''} a__ : int ={ '''vocab_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt''', }, '''emoji_file''': { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json''', }, } a__ : Dict ={ '''abeja/gpt-neox-japanese-2.7b''': 2_048, } def lowercase__ ( __lowercase : Tuple , __lowercase : Dict ) -> Any: """simple docstring""" with open(__lowercase , 'r' , encoding='utf-8' ) as f: __UpperCamelCase = json.loads(f.read() ) __UpperCamelCase = collections.OrderedDict() __UpperCamelCase = collections.OrderedDict() __UpperCamelCase = collections.OrderedDict() with open(__lowercase , 'r' , encoding='utf-8' ) as f: __UpperCamelCase = f.readlines() __UpperCamelCase = [[t.rstrip('\n' )] if (t == ',' or ',' not in t) else t.rstrip('\n' ).split(',' ) for t in token] for idx, b in enumerate(__lowercase ): __UpperCamelCase = b __UpperCamelCase = idx for wd in b: __UpperCamelCase = idx return vocab, raw_vocab, ids_to_tokens, emoji class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : List[str] =PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Dict =["input_ids", "attention_mask"] def __init__( self : Dict , __A : List[Any] , __A : Union[str, Any] , __A : List[str]="<|endoftext|>" , __A : int="<|endoftext|>" , __A : Optional[Any]="<|startoftext|>" , __A : Dict="<|endoftext|>" , __A : Dict=False , **__A : Any , ): super().__init__( unk_token=__A , pad_token=__A , bos_token=__A , eos_token=__A , do_clean_text=__A , **__A , ) if not os.path.isfile(__A ): raise ValueError( f'''Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained''' ' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) if not os.path.isfile(__A ): raise ValueError( f'''Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google''' ' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) __UpperCamelCase = do_clean_text __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = load_vocab_and_emoji(__A , __A ) __UpperCamelCase = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def _lowerCamelCase ( self : str ): # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def _lowerCamelCase ( self : Any ): return dict(self.raw_vocab , **self.added_tokens_encoder ) def _lowerCamelCase ( self : Optional[int] , __A : str ): return self.subword_tokenizer.tokenize(__A , clean=self.do_clean_text ) def _lowerCamelCase ( self : Dict , __A : int ): return self.vocab.get(__A , self.vocab.get(self.unk_token ) ) def _lowerCamelCase ( self : Any , __A : List[Any] ): return self.subword_tokenizer.convert_id_to_token(__A ) def _lowerCamelCase ( self : Tuple , __A : List[str] ): __UpperCamelCase = ''.join(__A ).strip() return out_string def _lowerCamelCase ( self : Union[str, Any] , __A : "Conversation" ): __UpperCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__A , add_special_tokens=__A ) + [self.eos_token_id] ) if len(__A ) > self.model_max_length: __UpperCamelCase = input_ids[-self.model_max_length :] return input_ids def _lowerCamelCase ( self : List[Any] , __A : str , __A : Optional[str] = None ): __UpperCamelCase = 0 if os.path.isdir(__A ): __UpperCamelCase = os.path.join( __A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join( __A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['emoji_file'] ) else: __UpperCamelCase = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(__A , 'w' , encoding='utf-8' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ' Please check that the vocabulary is not corrupted!' ) __UpperCamelCase = token_index writer.write(','.join(__A ) + '\n' ) index += 1 with open(__A , 'w' , encoding='utf-8' ) as writer: json.dump(self.emoji , __A ) return vocab_file, emoji_file class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Optional[Any] , __A : Optional[int] , __A : List[str] , __A : Dict ): __UpperCamelCase = vocab # same as swe __UpperCamelCase = ids_to_tokens # same as bpe __UpperCamelCase = emoji __UpperCamelCase = np.max([len(__A ) for w in self.vocab.keys()] ) __UpperCamelCase = re.compile(R'(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)' ) __UpperCamelCase = re.compile(R'[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*' ) __UpperCamelCase = re.compile(R'[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}' ) __UpperCamelCase = re.compile( R'([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) __UpperCamelCase = re.compile( R'(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) __UpperCamelCase = re.compile( R'((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*' ) __UpperCamelCase = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' __UpperCamelCase = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' __UpperCamelCase = str.maketrans({k: '<BLOCK>' for k in keisen + blocks} ) def __len__( self : List[Any] ): return len(self.ids_to_tokens ) def _lowerCamelCase ( self : List[str] , __A : List[str] ): __UpperCamelCase = self.content_repattera.sub('<URL>' , __A ) __UpperCamelCase = self.content_repattera.sub('<EMAIL>' , __A ) __UpperCamelCase = self.content_repattera.sub('<TEL>' , __A ) __UpperCamelCase = self.content_repattera.sub('<DATE>' , __A ) __UpperCamelCase = self.content_repattera.sub('<DATE>' , __A ) __UpperCamelCase = self.content_repattera.sub('<PRICE>' , __A ) __UpperCamelCase = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __UpperCamelCase = content.replace('<BLOCK><BLOCK>' , '<BLOCK>' ) return content def _lowerCamelCase ( self : List[Any] , __A : Any , __A : Union[str, Any]=False ): __UpperCamelCase = text.replace(' ' , '<SP>' ) __UpperCamelCase = text.replace(' ' , '<SP>' ) __UpperCamelCase = text.replace('\r\n' , '<BR>' ) __UpperCamelCase = text.replace('\n' , '<BR>' ) __UpperCamelCase = text.replace('\r' , '<BR>' ) __UpperCamelCase = text.replace('\t' , '<TAB>' ) __UpperCamelCase = text.replace('—' , 'ー' ) __UpperCamelCase = text.replace('−' , 'ー' ) for k, v in self.emoji["emoji"].items(): if k in text: __UpperCamelCase = text.replace(__A , __A ) if clean: __UpperCamelCase = self.clean_text(__A ) def check_simbol(__A : str ): __UpperCamelCase = x.encode() if len(__A ) == 1 and len(__A ) == 2: __UpperCamelCase = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xc2a1 and c <= 0xc2bf) or (c >= 0xc780 and c <= 0xc783) or (c >= 0xcab9 and c <= 0xcbbf) or (c >= 0xcc80 and c <= 0xcda2) ): return True return False def checkuae(__A : Tuple ): __UpperCamelCase = x.encode() if len(__A ) == 1 and len(__A ) == 3: __UpperCamelCase = (int(e[0] ) << 1_6) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xe2_8080 and c <= 0xe2_b07f: return True return False __UpperCamelCase = 0 __UpperCamelCase = [] while pos < len(__A ): __UpperCamelCase = min(len(__A ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 __UpperCamelCase = [] # (token_id, token, pos) for e in range(__A , __A , -1 ): __UpperCamelCase = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(__A ) > 2: __UpperCamelCase = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(__A ) > 0: # the smallest token_id is adopted __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = sorted(__A , key=lambda __A : x[0] )[0] result.append(__A ) __UpperCamelCase = e else: __UpperCamelCase = pos + 1 __UpperCamelCase = text[pos:end] if check_simbol(__A ): result.append('<KIGOU>' ) elif checkuae(__A ): result.append('<U2000U2BFF>' ) else: for i in wd.encode('utf-8' ): result.append('<|byte%d|>' % i ) __UpperCamelCase = end return result def _lowerCamelCase ( self : Union[str, Any] , __A : Any , __A : List[Any]="\n" ): __UpperCamelCase = [] __UpperCamelCase = [] __UpperCamelCase = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(__A ) > 0: words.append(bytearray(__A ).decode('utf-8' , errors='replace' ) ) __UpperCamelCase = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['emoji_inv'][word] ) elif word == "<SP>": words.append(' ' ) elif word == "<BR>": words.append(__A ) elif word == "<TAB>": words.append('\t' ) elif word == "<BLOCK>": words.append('▀' ) elif word == "<KIGOU>": words.append('ǀ' ) elif word == "<U2000U2BFF>": words.append('‖' ) else: words.append(__A ) if len(__A ) > 0: words.append(bytearray(__A ).decode('utf-8' , errors='replace' ) ) __UpperCamelCase = ''.join(__A ) return text
53
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
1
'''simple docstring''' from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize('repo_id' , ['canonical_dataset_name', 'org-name/dataset-name'] ) @pytest.mark.parametrize('path' , ['filename.csv', 'filename with blanks.csv'] ) @pytest.mark.parametrize('revision' , [None, 'v2'] ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : str ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = hf_hub_url(repo_id=__lowercase , path=__lowercase , revision=__lowercase ) assert url == F'''https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(__lowercase )}'''
53
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : List[Any] ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] =("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , __A : Optional[int] , __A : List[Any] ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : List[Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : List[Any] , *__A : Dict , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : List[Any] , *__A : List[str] , **__A : Dict ): return self.tokenizer.decode(*__A , **__A ) @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
1
'''simple docstring''' import math def lowercase__ ( __lowercase : int ) -> bool: """simple docstring""" return math.sqrt(__lowercase ) * math.sqrt(__lowercase ) == num def lowercase__ ( __lowercase : int ) -> bool: """simple docstring""" __UpperCamelCase = 0 __UpperCamelCase = n while left <= right: __UpperCamelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: __UpperCamelCase = mid - 1 else: __UpperCamelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
53
'''simple docstring''' from __future__ import annotations from typing import Any class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case : """simple docstring""" def __init__( self : List[Any] , __A : Any ): __UpperCamelCase = data __UpperCamelCase = None def __iter__( self : Optional[Any] ): __UpperCamelCase = self __UpperCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(__A ) yield node.data __UpperCamelCase = node.next_node @property def _lowerCamelCase ( self : List[str] ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": a__ : Dict =Node(1) a__ : Optional[int] =Node(2) a__ : List[str] =Node(3) a__ : Optional[int] =Node(4) print(root_node.has_loop) # False a__ : str =root_node.next_node print(root_node.has_loop) # True a__ : Optional[int] =Node(5) a__ : List[Any] =Node(6) a__ : int =Node(5) a__ : Tuple =Node(6) print(root_node.has_loop) # False a__ : str =Node(1) print(root_node.has_loop) # False
53
1
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging a__ : List[Any] =logging.get_logger(__name__) # TODO: upload to AWS a__ : Optional[int] ={ '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json''' ), } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] ="retribert" def __init__( self : int , __A : List[Any]=3_0_5_2_2 , __A : List[Any]=7_6_8 , __A : Union[str, Any]=8 , __A : Union[str, Any]=1_2 , __A : Optional[int]=3_0_7_2 , __A : int="gelu" , __A : List[str]=0.1 , __A : List[str]=0.1 , __A : Tuple=5_1_2 , __A : str=2 , __A : str=0.02 , __A : int=1e-12 , __A : Any=True , __A : Dict=1_2_8 , __A : Tuple=0 , **__A : int , ): super().__init__(pad_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = layer_norm_eps __UpperCamelCase = share_encoders __UpperCamelCase = projection_dim
53
'''simple docstring''' a__ : Optional[Any] =256 # Modulus to hash a string a__ : Dict =1_000_003 def lowercase__ ( __lowercase : str , __lowercase : str ) -> bool: """simple docstring""" __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) if p_len > t_len: return False __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 1 # Calculating the hash of pattern and substring of text for i in range(__lowercase ): __UpperCamelCase = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus __UpperCamelCase = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue __UpperCamelCase = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash __UpperCamelCase = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: """simple docstring""" __UpperCamelCase = 'abc1abc12' __UpperCamelCase = 'alskfjaldsabc1abc1abc12k23adsfabcabc' __UpperCamelCase = 'alskfjaldsk23adsfabcabc' assert rabin_karp(__lowercase , __lowercase ) and not rabin_karp(__lowercase , __lowercase ) # Test 2) __UpperCamelCase = 'ABABX' __UpperCamelCase = 'ABABZABABYABABX' assert rabin_karp(__lowercase , __lowercase ) # Test 3) __UpperCamelCase = 'AAAB' __UpperCamelCase = 'ABAAAAAB' assert rabin_karp(__lowercase , __lowercase ) # Test 4) __UpperCamelCase = 'abcdabcy' __UpperCamelCase = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(__lowercase , __lowercase ) # Test 5) __UpperCamelCase = 'Lü' __UpperCamelCase = 'Lüsai' assert rabin_karp(__lowercase , __lowercase ) __UpperCamelCase = 'Lue' assert not rabin_karp(__lowercase , __lowercase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
1
'''simple docstring''' from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
53
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class snake_case ( unittest.TestCase ): """simple docstring""" def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = 'ZinengTang/tvlt-base' __UpperCamelCase = tempfile.mkdtemp() def _lowerCamelCase ( self : Any , **__A : Any ): return TvltImageProcessor.from_pretrained(self.checkpoint , **__A ) def _lowerCamelCase ( self : Optional[Any] , **__A : Optional[int] ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **__A ) def _lowerCamelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.get_image_processor() __UpperCamelCase = self.get_feature_extractor() __UpperCamelCase = TvltProcessor(image_processor=__A , feature_extractor=__A ) processor.save_pretrained(self.tmpdirname ) __UpperCamelCase = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , __A ) self.assertIsInstance(processor.image_processor , __A ) def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_image_processor() __UpperCamelCase = self.get_feature_extractor() __UpperCamelCase = TvltProcessor(image_processor=__A , feature_extractor=__A ) __UpperCamelCase = np.ones([1_2_0_0_0] ) __UpperCamelCase = feature_extractor(__A , return_tensors='np' ) __UpperCamelCase = processor(audio=__A , return_tensors='np' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_image_processor() __UpperCamelCase = self.get_feature_extractor() __UpperCamelCase = TvltProcessor(image_processor=__A , feature_extractor=__A ) __UpperCamelCase = np.ones([3, 2_2_4, 2_2_4] ) __UpperCamelCase = image_processor(__A , return_tensors='np' ) __UpperCamelCase = processor(images=__A , return_tensors='np' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = self.get_image_processor() __UpperCamelCase = self.get_feature_extractor() __UpperCamelCase = TvltProcessor(image_processor=__A , feature_extractor=__A ) __UpperCamelCase = np.ones([1_2_0_0_0] ) __UpperCamelCase = np.ones([3, 2_2_4, 2_2_4] ) __UpperCamelCase = processor(audio=__A , images=__A ) self.assertListEqual(list(inputs.keys() ) , ['audio_values', 'audio_mask', 'pixel_values', 'pixel_mask'] ) # test if it raises when no input is passed with pytest.raises(__A ): processor() def _lowerCamelCase ( self : Union[str, Any] ): __UpperCamelCase = self.get_image_processor() __UpperCamelCase = self.get_feature_extractor() __UpperCamelCase = TvltProcessor(image_processor=__A , feature_extractor=__A ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='`processor` and `image_processor`+`feature_extractor` model input names do not match' , )
53
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
1
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 a__ : List[str] =get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a__ : Optional[Any] =get_tests_dir('''fixtures/vocab.json''') a__ : Optional[Any] =get_tests_dir('''fixtures''') class snake_case ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] =["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] def _lowerCamelCase ( self : Union[str, Any] ): __UpperCamelCase = 0 def _lowerCamelCase ( self : Dict ): __UpperCamelCase = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : int ): with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = WavaVecaConfig() __UpperCamelCase = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) # save in new folder model_config.save_pretrained(__A ) processor.save_pretrained(__A ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(__A , os.path.join(__A , __A ) ) copyfile(__A , os.path.join(__A , 'vocab.json' ) ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : Tuple ): with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = WavaVecaFeatureExtractor() __UpperCamelCase = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) __UpperCamelCase = WavaVecaProcessor(__A , __A ) # save in new folder processor.save_pretrained(__A ) # drop `processor_class` in tokenizer with open(os.path.join(__A , __A ) , 'r' ) as f: __UpperCamelCase = json.load(__A ) config_dict.pop('processor_class' ) with open(os.path.join(__A , __A ) , 'w' ) as f: f.write(json.dumps(__A ) ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : List[str] ): with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = WavaVecaFeatureExtractor() __UpperCamelCase = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) __UpperCamelCase = WavaVecaProcessor(__A , __A ) # save in new folder processor.save_pretrained(__A ) # drop `processor_class` in feature extractor with open(os.path.join(__A , __A ) , 'r' ) as f: __UpperCamelCase = json.load(__A ) config_dict.pop('processor_class' ) with open(os.path.join(__A , __A ) , 'w' ) as f: f.write(json.dumps(__A ) ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = WavaVecaConfig(processor_class='Wav2Vec2Processor' ) model_config.save_pretrained(__A ) # copy relevant files copyfile(__A , os.path.join(__A , 'vocab.json' ) ) # create emtpy sample processor with open(os.path.join(__A , __A ) , 'w' ) as f: f.write('{}' ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _lowerCamelCase ( self : int ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__A ): __UpperCamelCase = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(__A ): __UpperCamelCase = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=__A ) __UpperCamelCase = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' , trust_remote_code=__A ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) __UpperCamelCase = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) __UpperCamelCase = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version __UpperCamelCase = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=__A , use_fast=__A ) __UpperCamelCase = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def _lowerCamelCase ( self : Dict ): try: AutoConfig.register('custom' , __A ) AutoFeatureExtractor.register(__A , __A ) AutoTokenizer.register(__A , slow_tokenizer_class=__A ) AutoProcessor.register(__A , __A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoProcessor.register(__A , __A ) # Now that the config is registered, it can be used as any other config with the auto-API __UpperCamelCase = CustomFeatureExtractor.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase = os.path.join(__A , 'vocab.txt' ) with open(__A , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __UpperCamelCase = CustomTokenizer(__A ) __UpperCamelCase = CustomProcessor(__A , __A ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(__A ) __UpperCamelCase = AutoProcessor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowerCamelCase ( self : Optional[int] ): class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =False class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =False class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] ="AutoFeatureExtractor" SCREAMING_SNAKE_CASE_ : List[Any] ="AutoTokenizer" SCREAMING_SNAKE_CASE_ : Union[str, Any] =False try: AutoConfig.register('custom' , __A ) AutoFeatureExtractor.register(__A , __A ) AutoTokenizer.register(__A , slow_tokenizer_class=__A ) AutoProcessor.register(__A , __A ) # If remote code is not set, the default is to use local classes. __UpperCamelCase = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. __UpperCamelCase = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=__A ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. __UpperCamelCase = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=__A ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(processor.__class__.__name__ , 'BertTokenizerFast' ) def _lowerCamelCase ( self : int ): __UpperCamelCase = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-convnext' ) self.assertEqual(processor.__class__.__name__ , 'ConvNextImageProcessor' ) @is_staging_test class snake_case ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] =["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def _lowerCamelCase ( cls : Optional[Any] ): __UpperCamelCase = TOKEN HfFolder.save_token(__A ) @classmethod def _lowerCamelCase ( cls : int ): try: delete_repo(token=cls._token , repo_id='test-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-processor' ) except HTTPError: pass def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = WavaVecaProcessor.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(__A , 'test-processor' ) , push_to_hub=__A , use_auth_token=self._token ) __UpperCamelCase = WavaVecaProcessor.from_pretrained(f'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(__A , getattr(new_processor.feature_extractor , __A ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = WavaVecaProcessor.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(__A , 'test-processor-org' ) , push_to_hub=__A , use_auth_token=self._token , organization='valid_org' , ) __UpperCamelCase = WavaVecaProcessor.from_pretrained('valid_org/test-processor-org' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(__A , getattr(new_processor.feature_extractor , __A ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _lowerCamelCase ( self : Tuple ): CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() __UpperCamelCase = CustomFeatureExtractor.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase = os.path.join(__A , 'vocab.txt' ) with open(__A , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __UpperCamelCase = CustomTokenizer(__A ) __UpperCamelCase = CustomProcessor(__A , __A ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f'''{USER}/test-dynamic-processor''' , token=self._token ) __UpperCamelCase = Repository(__A , clone_from=f'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(__A ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { 'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor', 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(__A , 'tokenizer_config.json' ) ) as f: __UpperCamelCase = json.load(__A ) self.assertDictEqual( tokenizer_config['auto_map'] , { 'AutoTokenizer': ['custom_tokenization.CustomTokenizer', None], 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(__A , 'custom_feature_extraction.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(__A , 'custom_tokenization.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(__A , 'custom_processing.py' ) ) ) repo.push_to_hub() __UpperCamelCase = AutoProcessor.from_pretrained(f'''{USER}/test-dynamic-processor''' , trust_remote_code=__A ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , 'CustomProcessor' )
53
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
1
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers a__ : List[str] ='''3''' print('''Python version:''', sys.version) print('''transformers version:''', transformers.__version__) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) print('''NCCL version:''', torch.cuda.nccl.version()) except ImportError: print('''Torch version:''', None) try: import deepspeed print('''DeepSpeed version:''', deepspeed.__version__) except ImportError: print('''DeepSpeed version:''', None) try: import tensorflow as tf print('''TensorFlow version:''', tf.__version__) print('''TF GPUs available:''', bool(tf.config.list_physical_devices('''GPU'''))) print('''Number of TF GPUs available:''', len(tf.config.list_physical_devices('''GPU'''))) except ImportError: print('''TensorFlow version:''', None)
53
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def lowercase__ ( __lowercase : Tuple ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = emb.weight.shape __UpperCamelCase = nn.Linear(__lowercase , __lowercase , bias=__lowercase ) __UpperCamelCase = emb.weight.data return lin_layer def lowercase__ ( __lowercase : int , __lowercase : List[str]="facebook/mbart-large-en-ro" , __lowercase : str=False , __lowercase : List[Any]=False ) -> int: """simple docstring""" __UpperCamelCase = torch.load(__lowercase , map_location='cpu' )['model'] remove_ignore_keys_(__lowercase ) __UpperCamelCase = state_dict['encoder.embed_tokens.weight'].shape[0] __UpperCamelCase = MBartConfig.from_pretrained(__lowercase , vocab_size=__lowercase ) if mbart_aa and finetuned: __UpperCamelCase = 'relu' __UpperCamelCase = state_dict['decoder.embed_tokens.weight'] __UpperCamelCase = MBartForConditionalGeneration(__lowercase ) model.model.load_state_dict(__lowercase ) if finetuned: __UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": a__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') a__ : Union[str, Any] =parser.parse_args() a__ : str =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
53
1
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
'''simple docstring''' import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Any , __A : Dict , __A : str , __A : List[Any]=1_0_2_4 , __A : Tuple=1_0_2_4 , __A : str=3.6 ): __UpperCamelCase = tokenizer __UpperCamelCase = tokenizer.bos_token_id __UpperCamelCase = dataset __UpperCamelCase = seq_length __UpperCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self : Any ): __UpperCamelCase = iter(self.dataset ) __UpperCamelCase = True while more_examples: __UpperCamelCase , __UpperCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__A )['content'] ) buffer_len += len(buffer[-1] ) except StopIteration: __UpperCamelCase = False break __UpperCamelCase = tokenizer(__A , truncation=__A )['input_ids'] __UpperCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__A ) , self.seq_length ): __UpperCamelCase = all_token_ids[i : i + self.seq_length] if len(__A ) == self.seq_length: yield torch.tensor(__A ) def lowercase__ ( __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = {'streaming': True} __UpperCamelCase = load_dataset(args.dataset_name , split='train' , **__lowercase ) __UpperCamelCase = ConstantLengthDataset(__lowercase , __lowercase , seq_length=args.seq_length ) __UpperCamelCase = DataLoader(__lowercase , batch_size=args.batch_size ) return eval_dataloader def lowercase__ ( __lowercase : Tuple ) -> Optional[Any]: """simple docstring""" model.eval() __UpperCamelCase = [] for step, batch in enumerate(__lowercase ): with torch.no_grad(): __UpperCamelCase = model(__lowercase , labels=__lowercase ) __UpperCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(__lowercase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __UpperCamelCase = torch.mean(torch.cat(__lowercase ) ) try: __UpperCamelCase = torch.exp(__lowercase ) except OverflowError: __UpperCamelCase = float('inf' ) return loss.item(), perplexity.item() # Setup Accelerator a__ : int =Accelerator() # Parse configuration a__ : Dict =HfArgumentParser(EvaluationArguments) a__ : Union[str, Any] =parser.parse_args() set_seed(args.seed) # Logging a__ : List[Any] =logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer a__ : Union[str, Any] =AutoModelForCausalLM.from_pretrained(args.model_ckpt) a__ : List[Any] =AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a__ : Union[str, Any] =create_dataloader(args) # Prepare everything with our `accelerator`. a__ , a__ : List[str] =accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') a__ , a__ : Any =evaluate(args) logger.info(f'loss/eval: {eval_loss}, perplexity: {perplexity}')
53
1
'''simple docstring''' from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar a__ : List[str] =TypeVar('''T''') class snake_case ( Generic[T] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : deque[T] # Cache store of keys SCREAMING_SNAKE_CASE_ : set[T] # References of the keys in cache SCREAMING_SNAKE_CASE_ : int =10 # Maximum capacity of cache def __init__( self : Optional[int] , __A : int ): __UpperCamelCase = deque() __UpperCamelCase = set() if not n: __UpperCamelCase = sys.maxsize elif n < 0: raise ValueError('n should be an integer greater than 0.' ) else: __UpperCamelCase = n def _lowerCamelCase ( self : int , __A : T ): if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: __UpperCamelCase = self.dq_store.pop() self.key_reference.remove(__A ) else: self.dq_store.remove(__A ) self.dq_store.appendleft(__A ) self.key_reference.add(__A ) def _lowerCamelCase ( self : str ): for k in self.dq_store: print(__A ) def __repr__( self : int ): return f'''LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}''' if __name__ == "__main__": import doctest doctest.testmod() a__ : LRUCache[str | int] =LRUCache(4) lru_cache.refer('''A''') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('''A''') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
53
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging a__ : Any =logging.get_logger(__name__) a__ : Optional[Any] ={ '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict ="gpt_neo" SCREAMING_SNAKE_CASE_ : Optional[int] =["past_key_values"] SCREAMING_SNAKE_CASE_ : List[Any] ={"num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self : Union[str, Any] , __A : Union[str, Any]=5_0_2_5_7 , __A : Any=2_0_4_8 , __A : Optional[Any]=2_0_4_8 , __A : Any=2_4 , __A : Union[str, Any]=[[["global", "local"], 1_2]] , __A : str=1_6 , __A : Optional[int]=None , __A : Union[str, Any]=2_5_6 , __A : Any="gelu_new" , __A : Dict=0.0 , __A : Optional[int]=0.0 , __A : int=0.0 , __A : List[str]=0.1 , __A : Any=1e-5 , __A : int=0.02 , __A : List[str]=True , __A : Tuple=5_0_2_5_6 , __A : Optional[Any]=5_0_2_5_6 , **__A : Optional[Any] , ): __UpperCamelCase = vocab_size __UpperCamelCase = max_position_embeddings __UpperCamelCase = hidden_size __UpperCamelCase = num_layers __UpperCamelCase = num_heads __UpperCamelCase = intermediate_size __UpperCamelCase = window_size __UpperCamelCase = activation_function __UpperCamelCase = resid_dropout __UpperCamelCase = embed_dropout __UpperCamelCase = attention_dropout __UpperCamelCase = classifier_dropout __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = initializer_range __UpperCamelCase = use_cache __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id __UpperCamelCase = attention_types __UpperCamelCase = self.expand_attention_types_params(__A ) if len(self.attention_layers ) != self.num_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.attention_layers)` == `config.num_layers` ' f'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' f'''`config.num_layers = {self.num_layers}`. ''' '`config.attention_layers` is prepared using `config.attention_types`. ' 'Please verify the value of `config.attention_types` argument.' ) super().__init__(bos_token_id=__A , eos_token_id=__A , **__A ) @staticmethod def _lowerCamelCase ( __A : Tuple ): __UpperCamelCase = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase__ ( __lowercase : Tuple , __lowercase : Any , __lowercase : Union[str, Any] , __lowercase : List[str] ) -> Any: """simple docstring""" import torch __UpperCamelCase = input.size() __UpperCamelCase = len(__lowercase ) __UpperCamelCase = shape[dimension] __UpperCamelCase = torch.arange(0 , __lowercase , __lowercase ) __UpperCamelCase = torch.div(sizedim - size , __lowercase , rounding_mode='floor' ) + 1 __UpperCamelCase = torch.arange(__lowercase ) + low_indices[:min_length][:, None] __UpperCamelCase = [slice(__lowercase )] * rank __UpperCamelCase = indices __UpperCamelCase = input[s] __UpperCamelCase = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" import torch __UpperCamelCase = torch.arange(1 , __lowercase ) __UpperCamelCase = torch.remainder(__lowercase , __lowercase ) __UpperCamelCase = remainders == 0 __UpperCamelCase = candidates[divisor_indices] __UpperCamelCase = torch.max(__lowercase ) return largest_divisor, torch.div(__lowercase , __lowercase , rounding_mode='floor' ) class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(__A , direction='inputs' ) __UpperCamelCase = {0: 'batch', 1: 'past_sequence + sequence'} else: __UpperCamelCase = {0: 'batch', 1: 'sequence'} return common_inputs @property def _lowerCamelCase ( self : int ): return self._config.num_heads def _lowerCamelCase ( self : List[str] , __A : PreTrainedTokenizer , __A : int = -1 , __A : int = -1 , __A : bool = False , __A : Optional[TensorType] = None , ): __UpperCamelCase = super(__A , self ).generate_dummy_inputs( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) # We need to order the input in the way they appears in the forward() __UpperCamelCase = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __UpperCamelCase , __UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values __UpperCamelCase = seqlen + 2 __UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __UpperCamelCase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(self.num_layers ) ] __UpperCamelCase = common_inputs['attention_mask'] if self.use_past: __UpperCamelCase = ordered_inputs['attention_mask'].dtype __UpperCamelCase = torch.cat( [ordered_inputs['attention_mask'], torch.ones(__A , __A , dtype=__A )] , dim=1 ) return ordered_inputs @property def _lowerCamelCase ( self : Dict ): return 1_3
53
1
'''simple docstring''' import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def lowercase__ ( __lowercase : Any ) -> Tuple: """simple docstring""" __UpperCamelCase = [] embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', F'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', F'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', F'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', F'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def lowercase__ ( __lowercase : Tuple , __lowercase : Dict ) -> Tuple: """simple docstring""" __UpperCamelCase = [] attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', F'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', F'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', F'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', F'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def lowercase__ ( __lowercase : List[Any] ) -> Dict: """simple docstring""" __UpperCamelCase = [] token.append((F'''cvt.encoder.stages.{idx}.cls_token''', 'stage2.cls_token') ) return token def lowercase__ ( ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def lowercase__ ( __lowercase : str , __lowercase : Tuple , __lowercase : Any , __lowercase : Any ) -> Dict: """simple docstring""" __UpperCamelCase = 'imagenet-1k-id2label.json' __UpperCamelCase = 1000 __UpperCamelCase = 'huggingface/label-files' __UpperCamelCase = num_labels __UpperCamelCase = json.load(open(cached_download(hf_hub_url(__lowercase , __lowercase , repo_type='dataset' ) ) , 'r' ) ) __UpperCamelCase = {int(__lowercase ): v for k, v in idalabel.items()} __UpperCamelCase = idalabel __UpperCamelCase = {v: k for k, v in idalabel.items()} __UpperCamelCase = __UpperCamelCase = CvtConfig(num_labels=__lowercase , idalabel=__lowercase , labelaid=__lowercase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": __UpperCamelCase = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": __UpperCamelCase = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: __UpperCamelCase = [2, 2, 20] __UpperCamelCase = [3, 12, 16] __UpperCamelCase = [192, 768, 1024] __UpperCamelCase = CvtForImageClassification(__lowercase ) __UpperCamelCase = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) __UpperCamelCase = image_size __UpperCamelCase = torch.load(__lowercase , map_location=torch.device('cpu' ) ) __UpperCamelCase = OrderedDict() __UpperCamelCase = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: __UpperCamelCase = list_of_state_dict + cls_token(__lowercase ) __UpperCamelCase = list_of_state_dict + embeddings(__lowercase ) for cnt in range(config.depth[idx] ): __UpperCamelCase = list_of_state_dict + attention(__lowercase , __lowercase ) __UpperCamelCase = list_of_state_dict + final() for gg in list_of_state_dict: print(__lowercase ) for i in range(len(__lowercase ) ): __UpperCamelCase = original_weights[list_of_state_dict[i][1]] model.load_state_dict(__lowercase ) model.save_pretrained(__lowercase ) image_processor.save_pretrained(__lowercase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": a__ : List[str] =argparse.ArgumentParser() parser.add_argument( '''--cvt_model''', default='''cvt-w24''', type=str, help='''Name of the cvt model you\'d like to convert.''', ) parser.add_argument( '''--image_size''', default=384, type=int, help='''Input Image Size''', ) parser.add_argument( '''--cvt_file_name''', default=r'''cvtmodels\CvT-w24-384x384-IN-22k.pth''', type=str, help='''Input Image Size''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) a__ : Dict =parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
53
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="naver-clova-ix/donut-base-finetuned-docvqa" SCREAMING_SNAKE_CASE_ : Dict =( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) SCREAMING_SNAKE_CASE_ : List[str] ="document_qa" SCREAMING_SNAKE_CASE_ : Union[str, Any] =AutoProcessor SCREAMING_SNAKE_CASE_ : Union[str, Any] =VisionEncoderDecoderModel SCREAMING_SNAKE_CASE_ : List[Any] =["image", "text"] SCREAMING_SNAKE_CASE_ : Any =["text"] def __init__( self : Optional[int] , *__A : List[str] , **__A : List[Any] ): if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.' ) super().__init__(*__A , **__A ) def _lowerCamelCase ( self : Any , __A : "Image" , __A : str ): __UpperCamelCase = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' __UpperCamelCase = task_prompt.replace('{user_input}' , __A ) __UpperCamelCase = self.pre_processor.tokenizer( __A , add_special_tokens=__A , return_tensors='pt' ).input_ids __UpperCamelCase = self.pre_processor(__A , return_tensors='pt' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _lowerCamelCase ( self : Union[str, Any] , __A : Optional[Any] ): return self.model.generate( inputs['pixel_values'].to(self.device ) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__A , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__A , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__A , ).sequences def _lowerCamelCase ( self : Tuple , __A : List[Any] ): __UpperCamelCase = self.pre_processor.batch_decode(__A )[0] __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '' ) __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '' ) __UpperCamelCase = re.sub(R'<.*?>' , '' , __A , count=1 ).strip() # remove first task start token __UpperCamelCase = self.pre_processor.tokenajson(__A ) return sequence["answer"]
53
1
'''simple docstring''' def lowercase__ ( __lowercase : int ) -> bool: """simple docstring""" return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('''Program to check whether a number is a Perfect number or not...''') a__ : List[str] =int(input('''Enter number: ''').strip()) print(f'{number} is {"" if perfect(number) else "not "}a Perfect Number.')
53
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
53
1
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class snake_case ( __lowerCamelCase , __lowerCamelCase ): """simple docstring""" @register_to_config def __init__( self : Any , __A : int = 7_6_8 , ): super().__init__() __UpperCamelCase = nn.Parameter(torch.zeros(1 , __A ) ) __UpperCamelCase = nn.Parameter(torch.ones(1 , __A ) ) def _lowerCamelCase ( self : Optional[Any] , __A : Optional[Union[str, torch.device]] = None , __A : Optional[torch.dtype] = None , ): __UpperCamelCase = nn.Parameter(self.mean.to(__A ).to(__A ) ) __UpperCamelCase = nn.Parameter(self.std.to(__A ).to(__A ) ) return self def _lowerCamelCase ( self : Optional[int] , __A : Optional[int] ): __UpperCamelCase = (embeds - self.mean) * 1.0 / self.std return embeds def _lowerCamelCase ( self : Optional[int] , __A : Tuple ): __UpperCamelCase = (embeds * self.std) + self.mean return embeds
53
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowercase__ ( __lowercase : Features ) -> Optional[int]: """simple docstring""" __UpperCamelCase = np.inf def set_batch_size(__lowercase : FeatureType ) -> None: nonlocal batch_size if isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__lowercase , __lowercase ) and feature.dtype == "binary": __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__lowercase , __lowercase ) return None if batch_size is np.inf else batch_size class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , __A : NestedDataStructureLike[PathLike] , __A : Optional[NamedSplit] = None , __A : Optional[Features] = None , __A : str = None , __A : bool = False , __A : bool = False , __A : Optional[int] = None , **__A : Dict , ): super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) __UpperCamelCase = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} __UpperCamelCase = _PACKAGED_DATASETS_MODULES['parquet'][1] __UpperCamelCase = Parquet( cache_dir=__A , data_files=__A , features=__A , hash=__A , **__A , ) def _lowerCamelCase ( self : Optional[int] ): # Build iterable dataset if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Dataset , __A : Union[PathLike, BinaryIO] , __A : Optional[int] = None , **__A : Dict , ): __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size or get_writer_batch_size(dataset.features ) __UpperCamelCase = parquet_writer_kwargs def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , 'wb+' ) as buffer: __UpperCamelCase = self._write(file_obj=__A , batch_size=__A , **self.parquet_writer_kwargs ) else: __UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__A , **self.parquet_writer_kwargs ) return written def _lowerCamelCase ( self : List[str] , __A : BinaryIO , __A : int , **__A : List[str] ): __UpperCamelCase = 0 __UpperCamelCase = parquet_writer_kwargs.pop('path_or_buf' , __A ) __UpperCamelCase = self.dataset.features.arrow_schema __UpperCamelCase = pq.ParquetWriter(__A , schema=__A , **__A ) for offset in logging.tqdm( range(0 , len(self.dataset ) , __A ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating parquet from Arrow format' , ): __UpperCamelCase = query_table( table=self.dataset._data , key=slice(__A , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__A ) written += batch.nbytes writer.close() return written
53
1
import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() UpperCAmelCase__ = logging.get_logger(__name__) def _a ( a :Union[str, Any] , a :List[str] , a :List[str] ) -> int: a = UniSpeechSatForSequenceClassification.from_pretrained(a , config=a ) a = downstream_dict['''projector.weight'''] a = downstream_dict['''projector.bias'''] a = downstream_dict['''model.post_net.linear.weight'''] a = downstream_dict['''model.post_net.linear.bias'''] return model def _a ( a :Any , a :str , a :List[Any] ) -> Tuple: a = UniSpeechSatForAudioFrameClassification.from_pretrained(a , config=a ) a = downstream_dict['''model.linear.weight'''] a = downstream_dict['''model.linear.bias'''] return model def _a ( a :Dict , a :int , a :Union[str, Any] ) -> Dict: a = UniSpeechSatForXVector.from_pretrained(a , config=a ) a = downstream_dict['''connector.weight'''] a = downstream_dict['''connector.bias'''] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): a = downstream_dict[ F"""model.framelevel_feature_extractor.module.{i}.kernel.weight""" ] a = downstream_dict[F"""model.framelevel_feature_extractor.module.{i}.kernel.bias"""] a = downstream_dict['''model.utterancelevel_feature_extractor.linear1.weight'''] a = downstream_dict['''model.utterancelevel_feature_extractor.linear1.bias'''] a = downstream_dict['''model.utterancelevel_feature_extractor.linear2.weight'''] a = downstream_dict['''model.utterancelevel_feature_extractor.linear2.bias'''] a = downstream_dict['''objective.W'''] return model @torch.no_grad() def _a ( a :str , a :List[Any] , a :Optional[int] , a :Any ) -> str: a = torch.load(a , map_location='''cpu''' ) a = checkpoint['''Downstream'''] a = UniSpeechSatConfig.from_pretrained(a ) a = WavaVecaFeatureExtractor.from_pretrained( a , return_attention_mask=a , do_normalize=a ) a = hf_config.architectures[0] if arch.endswith('''ForSequenceClassification''' ): a = convert_classification(a , a , a ) elif arch.endswith('''ForAudioFrameClassification''' ): a = convert_diarization(a , a , a ) elif arch.endswith('''ForXVector''' ): a = convert_xvector(a , a , a ) else: raise NotImplementedError(F"""S3PRL weights conversion is not supported for {arch}""" ) if hf_config.use_weighted_layer_sum: a = checkpoint['''Featurizer''']['''weights'''] hf_feature_extractor.save_pretrained(a ) hf_model.save_pretrained(a ) if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() parser.add_argument( "--base_model_name", default=None, type=str, help="Name of the huggingface pretrained base model." ) parser.add_argument("--config_path", default=None, type=str, help="Path to the huggingface classifier config.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to the s3prl checkpoint.") parser.add_argument("--model_dump_path", default=None, type=str, help="Path to the final converted model.") UpperCAmelCase__ = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
0
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def lowercase__ ( __lowercase : SplitDict ) -> int: """simple docstring""" __UpperCamelCase = split_dict._to_yaml_list() assert len(__lowercase ) == len(__lowercase ) __UpperCamelCase = SplitDict._from_yaml_list(__lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __UpperCamelCase = None # the split name of split_dict takes over the name of the split info object __UpperCamelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=__lowercase ), SplitInfo(dataset_name='my_dataset' )] ) def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" __UpperCamelCase = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
53
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE_: Dict ={ 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConfig', 'SqueezeBertOnnxConfig', ], 'tokenization_squeezebert': ['SqueezeBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: Optional[int] =['SqueezeBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: str =[ 'SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'SqueezeBertForMaskedLM', 'SqueezeBertForMultipleChoice', 'SqueezeBertForQuestionAnswering', 'SqueezeBertForSequenceClassification', 'SqueezeBertForTokenClassification', 'SqueezeBertModel', 'SqueezeBertModule', 'SqueezeBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_: List[Any] =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[str] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Any =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys a__ : str =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class __lowerCAmelCase (unittest.TestCase ): '''simple docstring''' def __init__(self : Dict , UpperCamelCase : str , UpperCamelCase : Union[str, Any]=7 , UpperCamelCase : Optional[int]=3 , UpperCamelCase : Optional[int]=18 , UpperCamelCase : List[Any]=30 , UpperCamelCase : Dict=400 , UpperCamelCase : Dict=True , UpperCamelCase : Optional[Any]=None , UpperCamelCase : List[Any]=True , UpperCamelCase : int=None , UpperCamelCase : Optional[int]=True , UpperCamelCase : Union[str, Any]=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , UpperCamelCase : Union[str, Any]=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , UpperCamelCase : Union[str, Any]=True , ): '''simple docstring''' lowercase__ = size if size is not None else {'''height''': 224, '''width''': 224} lowercase__ = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size lowercase__ = do_center_crop lowercase__ = crop_size lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std lowercase__ = do_convert_rgb def UpperCamelCase__ (self : Optional[Any] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def UpperCamelCase__ (self : List[str] , UpperCamelCase : Dict=False , UpperCamelCase : Optional[int]=False , UpperCamelCase : Tuple=False ): '''simple docstring''' assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: lowercase__ = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 255 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: lowercase__ = [] for i in range(self.batch_size ): lowercase__ ,lowercase__ = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(255 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension lowercase__ = [Image.fromarray(np.moveaxis(UpperCamelCase , 0 , -1 ) ) for x in image_inputs] if torchify: lowercase__ = [torch.from_numpy(UpperCamelCase ) for x in image_inputs] return image_inputs @require_torch @require_vision class __lowerCAmelCase (lowercase_ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ : str = ChineseCLIPImageProcessor if is_vision_available() else None def UpperCamelCase__ (self : Tuple ): '''simple docstring''' lowercase__ = ChineseCLIPImageProcessingTester(self , do_center_crop=UpperCamelCase ) @property def UpperCamelCase__ (self : List[str] ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ (self : str ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''size''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''image_std''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_convert_rgb''' ) ) def UpperCamelCase__ (self : Tuple ): '''simple docstring''' lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 224, '''width''': 224} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def UpperCamelCase__ (self : Dict ): '''simple docstring''' pass def UpperCamelCase__ (self : Any ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = self.image_processor_tester.prepare_inputs(equal_resolution=UpperCamelCase ) for image in image_inputs: self.assertIsInstance(UpperCamelCase , Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase__ = image_processing(UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ (self : Any ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = self.image_processor_tester.prepare_inputs(equal_resolution=UpperCamelCase , numpify=UpperCamelCase ) for image in image_inputs: self.assertIsInstance(UpperCamelCase , np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase__ = image_processing(UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ (self : List[Any] ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = self.image_processor_tester.prepare_inputs(equal_resolution=UpperCamelCase , torchify=UpperCamelCase ) for image in image_inputs: self.assertIsInstance(UpperCamelCase , torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase__ = image_processing(UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) @require_torch @require_vision class __lowerCAmelCase (lowercase_ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ : List[Any] = ChineseCLIPImageProcessor if is_vision_available() else None def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' lowercase__ = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=UpperCamelCase ) lowercase__ = 3 @property def UpperCamelCase__ (self : int ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ (self : int ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''size''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''center_crop''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''image_std''' ) ) self.assertTrue(hasattr(UpperCamelCase , '''do_convert_rgb''' ) ) def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' pass def UpperCamelCase__ (self : Tuple ): '''simple docstring''' lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = self.image_processor_tester.prepare_inputs(equal_resolution=UpperCamelCase ) for image in image_inputs: self.assertIsInstance(UpperCamelCase , Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched lowercase__ = image_processing(UpperCamelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
2
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a__ : str =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =["input_features", "attention_mask"] def __init__( self : Union[str, Any] , __A : Optional[int]=8_0 , __A : Tuple=1_6_0_0_0 , __A : Optional[Any]=8_0 , __A : Any=0.0 , __A : Any=True , __A : List[str]=True , __A : str=True , **__A : List[Any] , ): super().__init__(feature_size=__A , sampling_rate=__A , padding_value=__A , **__A ) __UpperCamelCase = num_mel_bins __UpperCamelCase = do_ceptral_normalize __UpperCamelCase = normalize_means __UpperCamelCase = normalize_vars __UpperCamelCase = True def _lowerCamelCase ( self : Union[str, Any] , __A : np.ndarray , ): __UpperCamelCase = waveform * (2**1_5) # Kaldi compliance: 16-bit signed integers __UpperCamelCase = torch.from_numpy(__A ).unsqueeze(0 ) __UpperCamelCase = ta_kaldi.fbank(__A , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( __A : np.ndarray , __A : int , __A : Optional[bool] = True , __A : Optional[bool] = True , __A : float = 0.0 , ): # make sure we normalize float32 arrays if normalize_means: __UpperCamelCase = x[:input_length].mean(axis=0 ) __UpperCamelCase = np.subtract(__A , __A ) if normalize_vars: __UpperCamelCase = x[:input_length].std(axis=0 ) __UpperCamelCase = np.divide(__A , __A ) if input_length < x.shape[0]: __UpperCamelCase = padding_value # make sure array is in float32 __UpperCamelCase = x.astype(np.floataa ) return x def _lowerCamelCase ( self : int , __A : List[np.ndarray] , __A : Optional[np.ndarray] = None ): __UpperCamelCase = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(__A , __A , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(__A , __A ) ] def __call__( self : List[Any] , __A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __A : Union[bool, str, PaddingStrategy] = False , __A : Optional[int] = None , __A : bool = False , __A : Optional[int] = None , __A : Optional[Union[str, TensorType]] = None , __A : Optional[int] = None , __A : Optional[bool] = None , **__A : Dict , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) __UpperCamelCase = isinstance(__A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __UpperCamelCase = is_batched_numpy or ( isinstance(__A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(__A , np.ndarray ): __UpperCamelCase = np.asarray(__A , dtype=np.floataa ) elif isinstance(__A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCamelCase = [raw_speech] # extract fbank features __UpperCamelCase = [self._extract_fbank_features(__A ) for waveform in raw_speech] # convert into correct format for padding __UpperCamelCase = BatchFeature({'input_features': features} ) __UpperCamelCase = self.pad( __A , padding=__A , max_length=__A , truncation=__A , pad_to_multiple_of=__A , return_attention_mask=__A , **__A , ) # make sure list is in array format __UpperCamelCase = padded_inputs.get('input_features' ) if isinstance(input_features[0] , __A ): __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for feature in input_features] __UpperCamelCase = padded_inputs.get('attention_mask' ) if attention_mask is not None: __UpperCamelCase = [np.asarray(__A , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __UpperCamelCase = ( np.array(__A , dtype=np.intaa ) if self._get_padding_strategies(__A , max_length=__A ) is not PaddingStrategy.DO_NOT_PAD else None ) __UpperCamelCase = self.normalize( padded_inputs['input_features'] , attention_mask=__A ) if return_tensors is not None: __UpperCamelCase = padded_inputs.convert_to_tensors(__A ) return padded_inputs
53
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''multiplicative_persistence() does not accept negative values''' ) A : Any = 0 A : List[Any] = str(snake_case__ ) while len(snake_case__ ) != 1: A : str = [int(snake_case__ ) for i in num_string] A : List[str] = 1 for i in range(0 , len(snake_case__ ) ): total *= numbers[i] A : Optional[Any] = str(snake_case__ ) steps += 1 return steps def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError('''additive_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''additive_persistence() does not accept negative values''' ) A : Tuple = 0 A : Dict = str(snake_case__ ) while len(snake_case__ ) != 1: A : Any = [int(snake_case__ ) for i in num_string] A : Optional[int] = 0 for i in range(0 , len(snake_case__ ) ): total += numbers[i] A : Tuple = str(snake_case__ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
3
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : List[Any] =logging.get_logger(__name__) a__ : List[Any] ={ '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_text_model" def __init__( self : str , __A : List[Any]=2_5_0_0_0_2 , __A : Any=1_0_2_4 , __A : int=2_4 , __A : Dict=1_6 , __A : Optional[Any]=4_0_9_6 , __A : Union[str, Any]="gelu" , __A : Dict=0.1 , __A : Dict=0.1 , __A : List[str]=5_1_4 , __A : Optional[int]=1 , __A : int=0.02 , __A : Optional[Any]=0.02 , __A : Optional[Any]=1e-05 , __A : Dict=1 , __A : List[Any]=0 , __A : int=2 , __A : Tuple="absolute" , __A : Optional[Any]=True , __A : Optional[int]=7_6_8 , **__A : List[str] , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = use_cache __UpperCamelCase = project_dim class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_vision_model" def __init__( self : List[Any] , __A : Union[str, Any]=7_6_8 , __A : Optional[int]=3_0_7_2 , __A : Optional[Any]=5_1_2 , __A : Tuple=1_2 , __A : Union[str, Any]=1_2 , __A : Optional[int]=3 , __A : Dict=2_2_4 , __A : Tuple=3_2 , __A : str="quick_gelu" , __A : Dict=1e-5 , __A : Optional[int]=0.0 , __A : List[Any]=0.02 , __A : int=1.0 , **__A : Optional[int] , ): super().__init__(**__A ) __UpperCamelCase = hidden_size __UpperCamelCase = intermediate_size __UpperCamelCase = projection_dim __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = num_channels __UpperCamelCase = patch_size __UpperCamelCase = image_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = attention_dropout __UpperCamelCase = layer_norm_eps __UpperCamelCase = hidden_act @classmethod def _lowerCamelCase ( cls : Optional[Any] , __A : Union[str, os.PathLike] , **__A : Optional[Any] ): cls._set_token_in_kwargs(__A ) __UpperCamelCase , __UpperCamelCase = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('model_type' ) == "altclip": __UpperCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] ="altclip" SCREAMING_SNAKE_CASE_ : Optional[int] =True def __init__( self : Any , __A : List[str]=None , __A : List[Any]=None , __A : List[str]=7_6_8 , __A : List[str]=2.6592 , **__A : Dict ): # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __UpperCamelCase = kwargs.pop('text_config_dict' , __A ) __UpperCamelCase = kwargs.pop('vision_config_dict' , __A ) super().__init__(**__A ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __UpperCamelCase = {} # This is the complete result when using `text_config_dict`. __UpperCamelCase = AltCLIPTextConfig(**__A ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `text_config_dict` and `text_config` but with different values. ''' f'''The value `text_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The ''' f'''value `text_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __UpperCamelCase = {} # This is the complete result when using `vision_config_dict`. __UpperCamelCase = AltCLIPVisionConfig(**__A ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __UpperCamelCase = { str(__A ): value for key, value in _vision_config_dict['id2label'].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `vision_config_dict` and `vision_config` but with different ''' f'''values. The value `vision_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. ''' f'''The value `vision_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __UpperCamelCase = {} logger.info('`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.' ) if vision_config is None: __UpperCamelCase = {} logger.info('`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.' ) __UpperCamelCase = AltCLIPTextConfig(**__A ) __UpperCamelCase = AltCLIPVisionConfig(**__A ) __UpperCamelCase = projection_dim __UpperCamelCase = logit_scale_init_value __UpperCamelCase = 1.0 @classmethod def _lowerCamelCase ( cls : Union[str, Any] , __A : AltCLIPTextConfig , __A : AltCLIPVisionConfig , **__A : Optional[Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) __UpperCamelCase = self.text_config.to_dict() __UpperCamelCase = self.vision_config.to_dict() __UpperCamelCase = self.__class__.model_type return output
53
0
'''simple docstring''' def a_ ( lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ): lowerCAmelCase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def a_ ( ): print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
4
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Any ) -> Optional[Any]: """simple docstring""" with open(__lowercase ) as metadata_file: __UpperCamelCase = json.load(__lowercase ) __UpperCamelCase = LukeConfig(use_entity_aware_attention=__lowercase , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __UpperCamelCase = torch.load(__lowercase , map_location='cpu' ) # Load the entity vocab file __UpperCamelCase = load_entity_vocab(__lowercase ) __UpperCamelCase = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __UpperCamelCase = AddedToken('<ent>' , lstrip=__lowercase , rstrip=__lowercase ) __UpperCamelCase = AddedToken('<ent2>' , lstrip=__lowercase , rstrip=__lowercase ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(__lowercase , __lowercase ) __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase ) # Initialize the embeddings of the special tokens __UpperCamelCase = state_dict['embeddings.word_embeddings.weight'] __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __UpperCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __UpperCamelCase = F'''encoder.layer.{layer_index}.attention.self.''' __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __UpperCamelCase = state_dict['entity_embeddings.entity_embeddings.weight'] __UpperCamelCase = entity_emb[entity_vocab['[MASK]']] __UpperCamelCase = LukeModel(config=__lowercase ).eval() __UpperCamelCase , __UpperCamelCase = model.load_state_dict(__lowercase , strict=__lowercase ) if not (len(__lowercase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {', '.join(__lowercase )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' F''' {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}''' ) # Check outputs __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase , task='entity_classification' ) __UpperCamelCase = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __UpperCamelCase = (39, 42) __UpperCamelCase = tokenizer(__lowercase , entity_spans=[span] , add_prefix_space=__lowercase , return_tensors='pt' ) __UpperCamelCase = model(**__lowercase ) # Verify word hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 42, 1024) ) __UpperCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base __UpperCamelCase = torch.Size((1, 42, 768) ) __UpperCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 1, 1024) ) __UpperCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base __UpperCamelCase = torch.Size((1, 1, 768) ) __UpperCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(__lowercase ) ) model.save_pretrained(__lowercase ) def lowercase__ ( __lowercase : Dict ) -> List[str]: """simple docstring""" __UpperCamelCase = {} with open(__lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(__lowercase ): __UpperCamelCase , __UpperCamelCase = line.rstrip().split('\t' ) __UpperCamelCase = index return entity_vocab if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : str =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
53
0
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast UpperCAmelCase__ = datasets.utils.logging.get_logger(__name__) @dataclass class lowerCamelCase__ ( datasets.BuilderConfig): SCREAMING_SNAKE_CASE__ = 10000 SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None class lowerCamelCase__ ( datasets.ArrowBasedBuilder): SCREAMING_SNAKE_CASE__ = ParquetConfig def __A (self ) -> Union[str, Any]: return datasets.DatasetInfo(features=self.config.features ) def __A (self , UpperCAmelCase ) -> List[str]: if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}" ) _lowercase =dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase , (str, list, tuple) ): _lowercase =data_files if isinstance(UpperCAmelCase , UpperCAmelCase ): _lowercase =[files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowercase =[dl_manager.iter_files(UpperCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] _lowercase =[] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase , UpperCAmelCase ): _lowercase =[files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowercase =[dl_manager.iter_files(UpperCAmelCase ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(UpperCAmelCase ): with open(UpperCAmelCase , '''rb''' ) as f: _lowercase =datasets.Features.from_arrow_schema(pq.read_schema(UpperCAmelCase ) ) break splits.append(datasets.SplitGenerator(name=UpperCAmelCase , gen_kwargs={'''files''': files} ) ) return splits def __A (self , UpperCAmelCase ) -> pa.Table: if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example _lowercase =table_cast(UpperCAmelCase , self.info.features.arrow_schema ) return pa_table def __A (self , UpperCAmelCase ) -> Tuple: _lowercase =self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( f"Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'" ) for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase ) ): with open(UpperCAmelCase , '''rb''' ) as f: _lowercase =pq.ParquetFile(UpperCAmelCase ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): _lowercase =pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield f"{file_idx}_{batch_idx}", self._cast_table(UpperCAmelCase ) except ValueError as e: logger.error(f"Failed to read file '{file}' with error {type(UpperCAmelCase )}: {e}" ) raise
5
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case ( __lowerCamelCase ): """simple docstring""" def _lowerCamelCase ( self : Any ): __UpperCamelCase = tempfile.mkdtemp() __UpperCamelCase = 8 # DPR tok __UpperCamelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Optional[int] ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Union[str, Any] ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def _lowerCamelCase ( self : str ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _lowerCamelCase ( self : Any , __A : bool ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __UpperCamelCase = os.path.join(self.tmpdirname , 'dataset' ) __UpperCamelCase = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _lowerCamelCase ( self : int ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __UpperCamelCase = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __UpperCamelCase = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __UpperCamelCase = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__A , open(__A , 'wb' ) ) __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_legacy_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __A ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Optional[Any] ): import torch __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) __UpperCamelCase = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors='pt' , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dpr_ctx_encoder_tokenizer() __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __A ) # check for doc token related keys in dictionary.
53
0
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def __lowerCAmelCase ( a__ , a__ , a__ ) -> Optional[int]: # Initialise PyTorch model __a = RemBertConfig.from_json_file(a__ ) print('''Building PyTorch model from configuration: {}'''.format(str(a__ ) ) ) __a = RemBertModel(a__ ) # Load weights from tf checkpoint load_tf_weights_in_rembert(a__ , a__ , a__ ) # Save pytorch-model print('''Save PyTorch model to {}'''.format(a__ ) ) torch.save(model.state_dict() , a__ ) if __name__ == "__main__": A : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--rembert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained RemBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A : Tuple = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
6
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[Any] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys a__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
7
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: lowerCAmelCase_ = None lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase_ = { '''vocab_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/spiece.model''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/spiece.model''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/spiece.model''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''albert-base-v1''': '''https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json''', '''albert-large-v1''': '''https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json''', '''albert-xlarge-v1''': '''https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json''', '''albert-xxlarge-v1''': '''https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json''', '''albert-base-v2''': '''https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json''', '''albert-large-v2''': '''https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json''', '''albert-xlarge-v2''': '''https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json''', '''albert-xxlarge-v2''': '''https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json''', }, } lowerCAmelCase_ = { '''albert-base-v1''': 5_12, '''albert-large-v1''': 5_12, '''albert-xlarge-v1''': 5_12, '''albert-xxlarge-v1''': 5_12, '''albert-base-v2''': 5_12, '''albert-large-v2''': 5_12, '''albert-xlarge-v2''': 5_12, '''albert-xxlarge-v2''': 5_12, } lowerCAmelCase_ = '''▁''' class snake_case_ ( __A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : List[Any] = AlbertTokenizer def __init__( self : Tuple , _UpperCamelCase : Any=None , _UpperCamelCase : str=None , _UpperCamelCase : Tuple=True , _UpperCamelCase : Optional[Any]=True , _UpperCamelCase : Union[str, Any]=False , _UpperCamelCase : str="[CLS]" , _UpperCamelCase : Dict="[SEP]" , _UpperCamelCase : Tuple="<unk>" , _UpperCamelCase : str="[SEP]" , _UpperCamelCase : Optional[int]="<pad>" , _UpperCamelCase : Tuple="[CLS]" , _UpperCamelCase : int="[MASK]" , **_UpperCamelCase : Tuple , ) ->Tuple: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. snake_case_ = ( AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase , normalized=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else mask_token ) super().__init__( _UpperCamelCase , tokenizer_file=_UpperCamelCase , do_lower_case=_UpperCamelCase , remove_space=_UpperCamelCase , keep_accents=_UpperCamelCase , bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , unk_token=_UpperCamelCase , sep_token=_UpperCamelCase , pad_token=_UpperCamelCase , cls_token=_UpperCamelCase , mask_token=_UpperCamelCase , **_UpperCamelCase , ) snake_case_ = do_lower_case snake_case_ = remove_space snake_case_ = keep_accents snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True def snake_case__( self : Tuple , _UpperCamelCase : List[int] , _UpperCamelCase : Optional[List[int]] = None ) ->List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def snake_case__( self : str , _UpperCamelCase : List[int] , _UpperCamelCase : Optional[List[int]] = None ) ->List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__( self : str , _UpperCamelCase : str , _UpperCamelCase : Optional[str] = None ) ->Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_UpperCamelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return snake_case_ = os.path.join( _UpperCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCamelCase ): copyfile(self.vocab_file , _UpperCamelCase ) return (out_vocab_file,)
8
'''simple docstring''' # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys a__ : Tuple ='''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
53
0
import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _lowercase ( A__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = DebertaTokenizer SCREAMING_SNAKE_CASE__ : Union[str, Any] = True SCREAMING_SNAKE_CASE__ : Optional[int] = DebertaTokenizerFast def __magic_name__( self :List[str] ) -> str: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''[UNK]''', ] __SCREAMING_SNAKE_CASE : Optional[int] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) __SCREAMING_SNAKE_CASE : int = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __SCREAMING_SNAKE_CASE : Union[str, Any] = {'''unk_token''': '''[UNK]'''} __SCREAMING_SNAKE_CASE : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(lowerCAmelCase__ ) ) def __magic_name__( self :Any , **lowerCAmelCase__ :str ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def __magic_name__( self :int , lowerCAmelCase__ :int ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE : Optional[int] = '''lower newer''' __SCREAMING_SNAKE_CASE : int = '''lower newer''' return input_text, output_text def __magic_name__( self :str ) -> Optional[int]: __SCREAMING_SNAKE_CASE : Optional[Any] = self.get_tokenizer() __SCREAMING_SNAKE_CASE : Tuple = '''lower newer''' __SCREAMING_SNAKE_CASE : str = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = tokens + [tokenizer.unk_token] __SCREAMING_SNAKE_CASE : Optional[Any] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __magic_name__( self :List[Any] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE : List[str] = self.get_tokenizer() __SCREAMING_SNAKE_CASE : List[str] = tokenizer('''Hello''' , '''World''' ) __SCREAMING_SNAKE_CASE : Dict = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd['''token_type_ids'''] , lowerCAmelCase__ ) @slow def __magic_name__( self :Optional[int] ) -> List[Any]: __SCREAMING_SNAKE_CASE : Union[str, Any] = self.tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) __SCREAMING_SNAKE_CASE : int = tokenizer.encode('''sequence builders''' , add_special_tokens=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = tokenizer.encode('''multi-sequence build''' , add_special_tokens=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.encode( '''sequence builders''' , add_special_tokens=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ , lowerCAmelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def __magic_name__( self :Any ) -> Dict: __SCREAMING_SNAKE_CASE : Tuple = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: __SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) __SCREAMING_SNAKE_CASE : Dict = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer(lowerCAmelCase__ , padding=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = [tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) for seq in encoding['''input_ids''']] # fmt: off __SCREAMING_SNAKE_CASE : List[Any] = { '''input_ids''': [ [1, 2_118, 11_126, 565, 35, 83, 25_191, 163, 18_854, 13, 12_156, 12, 16_101, 25_376, 13_807, 9, 22_205, 27_893, 1_635, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 2_118, 11_126, 565, 24_536, 80, 43_797, 4_878, 7_373, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 133, 78, 65, 16, 10, 3_724, 1_538, 33_183, 11_303, 43_797, 1_938, 4, 870, 24_165, 29_105, 5, 739, 32_644, 33_183, 11_303, 36_173, 88, 80, 650, 7_821, 45_940, 6, 52, 2_559, 5, 1_836, 9, 5, 7_397, 13_171, 31, 5, 1_836, 9, 32_644, 33_183, 11_303, 4, 2] ], '''token_type_ids''': [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], '''attention_mask''': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on __SCREAMING_SNAKE_CASE : List[str] = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] self.assertDictEqual(encoding.data , lowerCAmelCase__ ) for expected, decoded in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ )
9
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "facebook/vit-mae-base": "https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "vit_mae" def __init__(self : int , UpperCAmelCase_ : List[Any]=768 , UpperCAmelCase_ : Optional[Any]=12 , UpperCAmelCase_ : Tuple=12 , UpperCAmelCase_ : int=3_072 , UpperCAmelCase_ : Union[str, Any]="gelu" , UpperCAmelCase_ : Union[str, Any]=0.0 , UpperCAmelCase_ : Tuple=0.0 , UpperCAmelCase_ : Optional[Any]=0.02 , UpperCAmelCase_ : List[Any]=1E-1_2 , UpperCAmelCase_ : Any=224 , UpperCAmelCase_ : List[str]=16 , UpperCAmelCase_ : Optional[int]=3 , UpperCAmelCase_ : List[str]=True , UpperCAmelCase_ : Dict=16 , UpperCAmelCase_ : str=512 , UpperCAmelCase_ : List[Any]=8 , UpperCAmelCase_ : List[str]=2_048 , UpperCAmelCase_ : Dict=0.75 , UpperCAmelCase_ : Tuple=False , **UpperCAmelCase_ : str , ) ->List[str]: '''simple docstring''' super().__init__(**UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =hidden_size lowerCamelCase__: int =num_hidden_layers lowerCamelCase__: str =num_attention_heads lowerCamelCase__: int =intermediate_size lowerCamelCase__: str =hidden_act lowerCamelCase__: Optional[int] =hidden_dropout_prob lowerCamelCase__: int =attention_probs_dropout_prob lowerCamelCase__: Dict =initializer_range lowerCamelCase__: int =layer_norm_eps lowerCamelCase__: List[str] =image_size lowerCamelCase__: Optional[Any] =patch_size lowerCamelCase__: Any =num_channels lowerCamelCase__: Optional[Any] =qkv_bias lowerCamelCase__: List[str] =decoder_num_attention_heads lowerCamelCase__: List[Any] =decoder_hidden_size lowerCamelCase__: str =decoder_num_hidden_layers lowerCamelCase__: List[Any] =decoder_intermediate_size lowerCamelCase__: Optional[int] =mask_ratio lowerCamelCase__: Optional[int] =norm_pix_loss
10
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : List[Any] ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] =("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , __A : Optional[int] , __A : List[Any] ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : List[Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : List[Any] , *__A : Dict , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : List[Any] , *__A : List[str] , **__A : Dict ): return self.tokenizer.decode(*__A , **__A ) @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
0
def _UpperCAmelCase (UpperCamelCase__ : int = 600851475143 ): try: _A : List[str] = int(UpperCamelCase__ ) except (TypeError, ValueError): raise TypeError("Parameter n must be int or castable to int." ) if n <= 0: raise ValueError("Parameter n must be greater than or equal to one." ) _A : List[Any] = 1 _A : int = 2 while i * i <= n: while n % i == 0: _A : str = i n //= i i += 1 if n > 1: _A : Dict = n return int(UpperCamelCase__ ) if __name__ == "__main__": print(f"{solution() = }")
11
'''simple docstring''' from __future__ import annotations from typing import Any class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case : """simple docstring""" def __init__( self : List[Any] , __A : Any ): __UpperCamelCase = data __UpperCamelCase = None def __iter__( self : Optional[Any] ): __UpperCamelCase = self __UpperCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(__A ) yield node.data __UpperCamelCase = node.next_node @property def _lowerCamelCase ( self : List[str] ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": a__ : Dict =Node(1) a__ : Optional[int] =Node(2) a__ : List[str] =Node(3) a__ : Optional[int] =Node(4) print(root_node.has_loop) # False a__ : str =root_node.next_node print(root_node.has_loop) # True a__ : Optional[int] =Node(5) a__ : List[Any] =Node(6) a__ : int =Node(5) a__ : Tuple =Node(6) print(root_node.has_loop) # False a__ : str =Node(1) print(root_node.has_loop) # False
53
0
UpperCAmelCase_ = { "km/h": 1.0, "m/s": 3.6, "mph": 1.60_9344, "knot": 1.852, } UpperCAmelCase_ = { "km/h": 1.0, "m/s": 0.2_7777_7778, "mph": 0.6_2137_1192, "knot": 0.5_3995_6803, } def lowerCamelCase__ ( A__ : float , A__ : str , A__ : str ): '''simple docstring''' if unit_to not in speed_chart or unit_from not in speed_chart_inverse: __lowerCamelCase = ( f'Incorrect \'from_type\' or \'to_type\' value: {unit_from!r}, {unit_to!r}\n' f'Valid values are: {", ".join(A__ )}' ) raise ValueError(A__ ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
12
'''simple docstring''' a__ : Optional[Any] =256 # Modulus to hash a string a__ : Dict =1_000_003 def lowercase__ ( __lowercase : str , __lowercase : str ) -> bool: """simple docstring""" __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) if p_len > t_len: return False __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 1 # Calculating the hash of pattern and substring of text for i in range(__lowercase ): __UpperCamelCase = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus __UpperCamelCase = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue __UpperCamelCase = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash __UpperCamelCase = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: """simple docstring""" __UpperCamelCase = 'abc1abc12' __UpperCamelCase = 'alskfjaldsabc1abc1abc12k23adsfabcabc' __UpperCamelCase = 'alskfjaldsk23adsfabcabc' assert rabin_karp(__lowercase , __lowercase ) and not rabin_karp(__lowercase , __lowercase ) # Test 2) __UpperCamelCase = 'ABABX' __UpperCamelCase = 'ABABZABABYABABX' assert rabin_karp(__lowercase , __lowercase ) # Test 3) __UpperCamelCase = 'AAAB' __UpperCamelCase = 'ABAAAAAB' assert rabin_karp(__lowercase , __lowercase ) # Test 4) __UpperCamelCase = 'abcdabcy' __UpperCamelCase = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(__lowercase , __lowercase ) # Test 5) __UpperCamelCase = 'Lü' __UpperCamelCase = 'Lüsai' assert rabin_karp(__lowercase , __lowercase ) __UpperCamelCase = 'Lue' assert not rabin_karp(__lowercase , __lowercase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : str = logging.get_logger(__name__) class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Any = '''timm_backbone''' def __init__( self : str , lowerCAmelCase__ : Any=None , lowerCAmelCase__ : Optional[Any]=3 , lowerCAmelCase__ : Any=True , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : int=None , **lowerCAmelCase__ : Optional[Any] , ): super().__init__(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = backbone SCREAMING_SNAKE_CASE_: Dict = num_channels SCREAMING_SNAKE_CASE_: Optional[Any] = features_only SCREAMING_SNAKE_CASE_: Optional[Any] = use_pretrained_backbone SCREAMING_SNAKE_CASE_: List[Any] = True SCREAMING_SNAKE_CASE_: List[str] = out_indices if out_indices is not None else (-1,)
13
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
0
from __future__ import annotations import queue class UpperCamelCase_ : '''simple docstring''' def __init__( self : str , UpperCAmelCase__ : str) ->Tuple: '''simple docstring''' A__ = data A__ = None A__ = None def SCREAMING_SNAKE_CASE ( ) -> TreeNode: """simple docstring""" print('''\n********Press N to stop entering at any point of time********\n''' ) A__ = input('''Enter the value of the root node: ''' ).strip().lower() A__ = queue.Queue() A__ = TreeNode(int(lowercase_ ) ) q.put(lowercase_ ) while not q.empty(): A__ = q.get() A__ = f"""Enter the left node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = left_node q.put(lowercase_ ) A__ = f"""Enter the right node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = right_node q.put(lowercase_ ) raise def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = [] while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(lowercase_ ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(lowercase_ ) A__ = n.left # end of while means current node doesn't have left child A__ = stack.pop() # start to traverse its right child A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: stack.append(lowercase_ ) A__ = n.left A__ = stack.pop() print(n.data , end=''',''' ) A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ , A__ = [], [] A__ = node stacka.append(lowercase_ ) while stacka: # to find the reversed order of post order, store it in stack2 A__ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(lowercase_ ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ = "" , lowercase_=50 , lowercase_="*" ) -> str: """simple docstring""" if not s: return "\n" + width * char A__ , A__ = divmod(width - len(lowercase_ ) - 2 , 2 ) return f"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt("""Binary Tree Traversals""")) _lowerCamelCase : TreeNode = build_tree() print(prompt("""Pre Order Traversal""")) pre_order(node) print(prompt() + """\n""") print(prompt("""In Order Traversal""")) in_order(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal""")) post_order(node) print(prompt() + """\n""") print(prompt("""Level Order Traversal""")) level_order(node) print(prompt() + """\n""") print(prompt("""Actual Level Order Traversal""")) level_order_actual(node) print("""*""" * 50 + """\n""") print(prompt("""Pre Order Traversal - Iteration Version""")) pre_order_iter(node) print(prompt() + """\n""") print(prompt("""In Order Traversal - Iteration Version""")) in_order_iter(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal - Iteration Version""")) post_order_iter(node) print(prompt())
14
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE :Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE :List[Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE :Dict = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE :Tuple = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE :Optional[Any] = ['LayoutLMv3FeatureExtractor'] SCREAMING_SNAKE_CASE :int = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys SCREAMING_SNAKE_CASE :Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
15
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
0
"""simple docstring""" from __future__ import annotations from collections import namedtuple def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> tuple: lowercase__ : Any = namedtuple('''result''' , '''name value''' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('''Only one argument must be 0''' ) elif power < 0: raise ValueError( '''Power cannot be negative in any electrical/electronics system''' ) elif voltage == 0: return result('''voltage''' , power / current ) elif current == 0: return result('''current''' , power / voltage ) elif power == 0: return result('''power''' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
16
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def lowercase__ ( __lowercase : Tuple ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = emb.weight.shape __UpperCamelCase = nn.Linear(__lowercase , __lowercase , bias=__lowercase ) __UpperCamelCase = emb.weight.data return lin_layer def lowercase__ ( __lowercase : int , __lowercase : List[str]="facebook/mbart-large-en-ro" , __lowercase : str=False , __lowercase : List[Any]=False ) -> int: """simple docstring""" __UpperCamelCase = torch.load(__lowercase , map_location='cpu' )['model'] remove_ignore_keys_(__lowercase ) __UpperCamelCase = state_dict['encoder.embed_tokens.weight'].shape[0] __UpperCamelCase = MBartConfig.from_pretrained(__lowercase , vocab_size=__lowercase ) if mbart_aa and finetuned: __UpperCamelCase = 'relu' __UpperCamelCase = state_dict['decoder.embed_tokens.weight'] __UpperCamelCase = MBartForConditionalGeneration(__lowercase ) model.model.load_state_dict(__lowercase ) if finetuned: __UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": a__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') a__ : Union[str, Any] =parser.parse_args() a__ : str =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
53
0
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class _lowerCAmelCase : """simple docstring""" def __init__( self : Any, UpperCAmelCase__ : List[Any], ): __lowercase = parent __lowercase = 1_3 __lowercase = 7 __lowercase = 3_0 __lowercase = self.seq_length + self.mem_len __lowercase = 1_5 __lowercase = True __lowercase = True __lowercase = 9_9 __lowercase = [1_0, 5_0, 8_0] __lowercase = 3_2 __lowercase = 3_2 __lowercase = 4 __lowercase = 8 __lowercase = 1_2_8 __lowercase = 2 __lowercase = 2 __lowercase = None __lowercase = 1 __lowercase = 0 __lowercase = 3 __lowercase = self.vocab_size - 1 __lowercase = 0.01 def _lowercase ( self : List[str] ): __lowercase = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) __lowercase = TransfoXLConfig( vocab_size=self.vocab_size, mem_len=self.mem_len, clamp_len=self.clamp_len, cutoffs=self.cutoffs, d_model=self.hidden_size, d_embed=self.d_embed, n_head=self.num_attention_heads, d_head=self.d_head, d_inner=self.d_inner, div_val=self.div_val, n_layer=self.num_hidden_layers, eos_token_id=self.eos_token_id, pad_token_id=self.vocab_size - 1, init_range=self.init_range, num_labels=self.num_labels, ) return (config, input_ids_a, input_ids_a, lm_labels) def _lowercase ( self : List[Any] ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _lowercase ( self : int, UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : int, UpperCAmelCase__ : Tuple ): __lowercase = TFTransfoXLModel(UpperCAmelCase__ ) __lowercase ,__lowercase = model(UpperCAmelCase__ ).to_tuple() __lowercase = {"input_ids": input_ids_a, "mems": mems_a} __lowercase ,__lowercase = model(UpperCAmelCase__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape, (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape, (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a], [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers, ) self.parent.assertListEqual( [mem.shape for mem in mems_a], [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers, ) def _lowercase ( self : List[str], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Dict, UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : List[str] ): __lowercase = TFTransfoXLLMHeadModel(UpperCAmelCase__ ) __lowercase ,__lowercase = model(UpperCAmelCase__ ).to_tuple() __lowercase = {"input_ids": input_ids_a, "labels": lm_labels} __lowercase ,__lowercase = model(UpperCAmelCase__ ).to_tuple() __lowercase ,__lowercase = model([input_ids_a, mems_a] ).to_tuple() __lowercase = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} __lowercase ,__lowercase = model(UpperCAmelCase__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape, (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a], [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers, ) self.parent.assertEqual(lm_logits_a.shape, (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a], [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers, ) def _lowercase ( self : Optional[Any], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : str, UpperCAmelCase__ : str, UpperCAmelCase__ : Optional[int] ): __lowercase = TFTransfoXLForSequenceClassification(UpperCAmelCase__ ) __lowercase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def _lowercase ( self : Any ): __lowercase = self.prepare_config_and_inputs() ((__lowercase) ,(__lowercase) ,(__lowercase) ,(__lowercase)) = config_and_inputs __lowercase = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class _lowerCAmelCase ( lowercase ,lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) __UpperCAmelCase : Union[str, Any] = () if is_tf_available() else () __UpperCAmelCase : List[str] = ( { "feature-extraction": TFTransfoXLModel, "text-classification": TFTransfoXLForSequenceClassification, "text-generation": TFTransfoXLLMHeadModel, "zero-shot": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented __UpperCAmelCase : Tuple = False __UpperCAmelCase : List[str] = False __UpperCAmelCase : Any = False __UpperCAmelCase : List[Any] = False def _lowercase ( self : str, UpperCAmelCase__ : List[str], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Any, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : int ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _lowercase ( self : List[str] ): __lowercase = TFTransfoXLModelTester(self ) __lowercase = ConfigTester(self, config_class=UpperCAmelCase__, d_embed=3_7 ) def _lowercase ( self : int ): self.config_tester.run_common_tests() def _lowercase ( self : Union[str, Any] ): self.model_tester.set_seed() __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*UpperCAmelCase__ ) def _lowercase ( self : Tuple ): self.model_tester.set_seed() __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*UpperCAmelCase__ ) def _lowercase ( self : int ): __lowercase ,__lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: __lowercase = model_class(UpperCAmelCase__ ) assert isinstance(model.get_input_embeddings(), tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: __lowercase = model.get_output_embeddings() assert isinstance(UpperCAmelCase__, tf.keras.layers.Layer ) __lowercase = model.get_bias() assert name is None else: __lowercase = model.get_output_embeddings() assert x is None __lowercase = model.get_bias() assert name is None def _lowercase ( self : List[str] ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _lowercase ( self : Tuple ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = TFTransfoXLModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss." ) def _lowercase ( self : Any ): pass @require_tf class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @unittest.skip("Skip test until #12651 is resolved." ) @slow def _lowercase ( self : List[str] ): __lowercase = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off __lowercase = tf.convert_to_tensor([[3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0]], dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off __lowercase = [3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0,3_3,1,1_8_5_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_8,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> __lowercase = model.generate(UpperCAmelCase__, max_length=2_0_0, do_sample=UpperCAmelCase__ ) self.assertListEqual(output_ids[0].numpy().tolist(), UpperCAmelCase__ )
17
'''simple docstring''' import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Any , __A : Dict , __A : str , __A : List[Any]=1_0_2_4 , __A : Tuple=1_0_2_4 , __A : str=3.6 ): __UpperCamelCase = tokenizer __UpperCamelCase = tokenizer.bos_token_id __UpperCamelCase = dataset __UpperCamelCase = seq_length __UpperCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self : Any ): __UpperCamelCase = iter(self.dataset ) __UpperCamelCase = True while more_examples: __UpperCamelCase , __UpperCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__A )['content'] ) buffer_len += len(buffer[-1] ) except StopIteration: __UpperCamelCase = False break __UpperCamelCase = tokenizer(__A , truncation=__A )['input_ids'] __UpperCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__A ) , self.seq_length ): __UpperCamelCase = all_token_ids[i : i + self.seq_length] if len(__A ) == self.seq_length: yield torch.tensor(__A ) def lowercase__ ( __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = {'streaming': True} __UpperCamelCase = load_dataset(args.dataset_name , split='train' , **__lowercase ) __UpperCamelCase = ConstantLengthDataset(__lowercase , __lowercase , seq_length=args.seq_length ) __UpperCamelCase = DataLoader(__lowercase , batch_size=args.batch_size ) return eval_dataloader def lowercase__ ( __lowercase : Tuple ) -> Optional[Any]: """simple docstring""" model.eval() __UpperCamelCase = [] for step, batch in enumerate(__lowercase ): with torch.no_grad(): __UpperCamelCase = model(__lowercase , labels=__lowercase ) __UpperCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(__lowercase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __UpperCamelCase = torch.mean(torch.cat(__lowercase ) ) try: __UpperCamelCase = torch.exp(__lowercase ) except OverflowError: __UpperCamelCase = float('inf' ) return loss.item(), perplexity.item() # Setup Accelerator a__ : int =Accelerator() # Parse configuration a__ : Dict =HfArgumentParser(EvaluationArguments) a__ : Union[str, Any] =parser.parse_args() set_seed(args.seed) # Logging a__ : List[Any] =logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer a__ : Union[str, Any] =AutoModelForCausalLM.from_pretrained(args.model_ckpt) a__ : List[Any] =AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a__ : Union[str, Any] =create_dataloader(args) # Prepare everything with our `accelerator`. a__ , a__ : List[str] =accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') a__ , a__ : Any =evaluate(args) logger.info(f'loss/eval: {eval_loss}, perplexity: {perplexity}')
53
0
__lowerCamelCase : List[Any] = 8.3144598 def _snake_case ( lowerCAmelCase : float , lowerCAmelCase : float ): """simple docstring""" if temperature < 0: raise Exception("Temperature cannot be less than 0 K" ) if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example __lowerCamelCase : Dict = 3_00 __lowerCamelCase : str = 28 __lowerCamelCase : Optional[Any] = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
18
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging a__ : Any =logging.get_logger(__name__) a__ : Optional[Any] ={ '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict ="gpt_neo" SCREAMING_SNAKE_CASE_ : Optional[int] =["past_key_values"] SCREAMING_SNAKE_CASE_ : List[Any] ={"num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self : Union[str, Any] , __A : Union[str, Any]=5_0_2_5_7 , __A : Any=2_0_4_8 , __A : Optional[Any]=2_0_4_8 , __A : Any=2_4 , __A : Union[str, Any]=[[["global", "local"], 1_2]] , __A : str=1_6 , __A : Optional[int]=None , __A : Union[str, Any]=2_5_6 , __A : Any="gelu_new" , __A : Dict=0.0 , __A : Optional[int]=0.0 , __A : int=0.0 , __A : List[str]=0.1 , __A : Any=1e-5 , __A : int=0.02 , __A : List[str]=True , __A : Tuple=5_0_2_5_6 , __A : Optional[Any]=5_0_2_5_6 , **__A : Optional[Any] , ): __UpperCamelCase = vocab_size __UpperCamelCase = max_position_embeddings __UpperCamelCase = hidden_size __UpperCamelCase = num_layers __UpperCamelCase = num_heads __UpperCamelCase = intermediate_size __UpperCamelCase = window_size __UpperCamelCase = activation_function __UpperCamelCase = resid_dropout __UpperCamelCase = embed_dropout __UpperCamelCase = attention_dropout __UpperCamelCase = classifier_dropout __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = initializer_range __UpperCamelCase = use_cache __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id __UpperCamelCase = attention_types __UpperCamelCase = self.expand_attention_types_params(__A ) if len(self.attention_layers ) != self.num_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.attention_layers)` == `config.num_layers` ' f'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' f'''`config.num_layers = {self.num_layers}`. ''' '`config.attention_layers` is prepared using `config.attention_types`. ' 'Please verify the value of `config.attention_types` argument.' ) super().__init__(bos_token_id=__A , eos_token_id=__A , **__A ) @staticmethod def _lowerCamelCase ( __A : Tuple ): __UpperCamelCase = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase__ ( __lowercase : Tuple , __lowercase : Any , __lowercase : Union[str, Any] , __lowercase : List[str] ) -> Any: """simple docstring""" import torch __UpperCamelCase = input.size() __UpperCamelCase = len(__lowercase ) __UpperCamelCase = shape[dimension] __UpperCamelCase = torch.arange(0 , __lowercase , __lowercase ) __UpperCamelCase = torch.div(sizedim - size , __lowercase , rounding_mode='floor' ) + 1 __UpperCamelCase = torch.arange(__lowercase ) + low_indices[:min_length][:, None] __UpperCamelCase = [slice(__lowercase )] * rank __UpperCamelCase = indices __UpperCamelCase = input[s] __UpperCamelCase = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" import torch __UpperCamelCase = torch.arange(1 , __lowercase ) __UpperCamelCase = torch.remainder(__lowercase , __lowercase ) __UpperCamelCase = remainders == 0 __UpperCamelCase = candidates[divisor_indices] __UpperCamelCase = torch.max(__lowercase ) return largest_divisor, torch.div(__lowercase , __lowercase , rounding_mode='floor' ) class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(__A , direction='inputs' ) __UpperCamelCase = {0: 'batch', 1: 'past_sequence + sequence'} else: __UpperCamelCase = {0: 'batch', 1: 'sequence'} return common_inputs @property def _lowerCamelCase ( self : int ): return self._config.num_heads def _lowerCamelCase ( self : List[str] , __A : PreTrainedTokenizer , __A : int = -1 , __A : int = -1 , __A : bool = False , __A : Optional[TensorType] = None , ): __UpperCamelCase = super(__A , self ).generate_dummy_inputs( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) # We need to order the input in the way they appears in the forward() __UpperCamelCase = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __UpperCamelCase , __UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values __UpperCamelCase = seqlen + 2 __UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __UpperCamelCase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(self.num_layers ) ] __UpperCamelCase = common_inputs['attention_mask'] if self.use_past: __UpperCamelCase = ordered_inputs['attention_mask'].dtype __UpperCamelCase = torch.cat( [ordered_inputs['attention_mask'], torch.ones(__A , __A , dtype=__A )] , dim=1 ) return ordered_inputs @property def _lowerCamelCase ( self : Dict ): return 1_3
53
0
def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = 0 while b > 0: if b & 1: lowerCamelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
19
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="naver-clova-ix/donut-base-finetuned-docvqa" SCREAMING_SNAKE_CASE_ : Dict =( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) SCREAMING_SNAKE_CASE_ : List[str] ="document_qa" SCREAMING_SNAKE_CASE_ : Union[str, Any] =AutoProcessor SCREAMING_SNAKE_CASE_ : Union[str, Any] =VisionEncoderDecoderModel SCREAMING_SNAKE_CASE_ : List[Any] =["image", "text"] SCREAMING_SNAKE_CASE_ : Any =["text"] def __init__( self : Optional[int] , *__A : List[str] , **__A : List[Any] ): if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.' ) super().__init__(*__A , **__A ) def _lowerCamelCase ( self : Any , __A : "Image" , __A : str ): __UpperCamelCase = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' __UpperCamelCase = task_prompt.replace('{user_input}' , __A ) __UpperCamelCase = self.pre_processor.tokenizer( __A , add_special_tokens=__A , return_tensors='pt' ).input_ids __UpperCamelCase = self.pre_processor(__A , return_tensors='pt' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _lowerCamelCase ( self : Union[str, Any] , __A : Optional[Any] ): return self.model.generate( inputs['pixel_values'].to(self.device ) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__A , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__A , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__A , ).sequences def _lowerCamelCase ( self : Tuple , __A : List[Any] ): __UpperCamelCase = self.pre_processor.batch_decode(__A )[0] __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '' ) __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '' ) __UpperCamelCase = re.sub(R'<.*?>' , '' , __A , count=1 ).strip() # remove first task start token __UpperCamelCase = self.pre_processor.tokenajson(__A ) return sequence["answer"]
53
0
from typing import Dict, Optional import numpy as np import datasets lowercase : List[str] = """ IoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union between the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation, the mean IoU of the image is calculated by taking the IoU of each class and averaging them. """ lowercase : Tuple = """ Args: predictions (`List[ndarray]`): List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. references (`List[ndarray]`): List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. num_labels (`int`): Number of classes (categories). ignore_index (`int`): Index that will be ignored during evaluation. nan_to_num (`int`, *optional*): If specified, NaN values will be replaced by the number defined by the user. label_map (`dict`, *optional*): If specified, dictionary mapping old label indices to new label indices. reduce_labels (`bool`, *optional*, defaults to `False`): Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background, and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255. Returns: `Dict[str, float | ndarray]` comprising various elements: - *mean_iou* (`float`): Mean Intersection-over-Union (IoU averaged over all categories). - *mean_accuracy* (`float`): Mean accuracy (averaged over all categories). - *overall_accuracy* (`float`): Overall accuracy on all images. - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`): Per category accuracy. - *per_category_iou* (`ndarray` of shape `(num_labels,)`): Per category IoU. Examples: >>> import numpy as np >>> mean_iou = datasets.load_metric(\"mean_iou\") >>> # suppose one has 3 different segmentation maps predicted >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]]) >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]]) >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]]) >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]]) >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]]) >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]]) >>> predicted = [predicted_1, predicted_2, predicted_3] >>> ground_truth = [actual_1, actual_2, actual_3] >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False) >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mean_iou': 0.47750000000000004, 'mean_accuracy': 0.5916666666666666, 'overall_accuracy': 0.5263157894736842, 'per_category_iou': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), 'per_category_accuracy': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])} """ lowercase : List[Any] = """\ @software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020, author = {{MMSegmentation Contributors}}, license = {Apache-2.0}, month = {7}, title = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}}, url = {https://github.com/open-mmlab/mmsegmentation}, year = {2020} }""" def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False , ) -> Union[str, Any]: if label_map is not None: for old_id, new_id in label_map.items(): lowercase : Union[str, Any] = new_id # turn into Numpy arrays lowercase : Optional[int] = np.array(SCREAMING_SNAKE_CASE__ ) lowercase : Optional[int] = np.array(SCREAMING_SNAKE_CASE__ ) if reduce_labels: lowercase : Optional[Any] = 255 lowercase : int = label - 1 lowercase : int = 255 lowercase : Tuple = label != ignore_index lowercase : Optional[int] = np.not_equal(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : int = pred_label[mask] lowercase : Dict = np.array(SCREAMING_SNAKE_CASE__ )[mask] lowercase : Union[str, Any] = pred_label[pred_label == label] lowercase : List[Any] = np.histogram(SCREAMING_SNAKE_CASE__ , bins=SCREAMING_SNAKE_CASE__ , range=(0, num_labels - 1) )[0] lowercase : Any = np.histogram(SCREAMING_SNAKE_CASE__ , bins=SCREAMING_SNAKE_CASE__ , range=(0, num_labels - 1) )[0] lowercase : Tuple = np.histogram(SCREAMING_SNAKE_CASE__ , bins=SCREAMING_SNAKE_CASE__ , range=(0, num_labels - 1) )[0] lowercase : str = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False , ) -> List[Any]: lowercase : Dict = np.zeros((num_labels,) , dtype=np.floataa ) lowercase : Dict = np.zeros((num_labels,) , dtype=np.floataa ) lowercase : Optional[Any] = np.zeros((num_labels,) , dtype=np.floataa ) lowercase : Optional[int] = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase , lowercase , lowercase , lowercase : List[str] = intersect_and_union( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False , ) -> Optional[Any]: lowercase , lowercase , lowercase , lowercase : Tuple = total_intersect_and_union( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # compute metrics lowercase : str = {} lowercase : List[Any] = total_area_intersect.sum() / total_area_label.sum() lowercase : List[Any] = total_area_intersect / total_area_union lowercase : Union[str, Any] = total_area_intersect / total_area_label lowercase : Union[str, Any] = np.nanmean(SCREAMING_SNAKE_CASE__ ) lowercase : List[str] = np.nanmean(SCREAMING_SNAKE_CASE__ ) lowercase : Tuple = all_acc lowercase : str = iou lowercase : int = acc if nan_to_num is not None: lowercase : Tuple = {metric: np.nan_to_num(SCREAMING_SNAKE_CASE__ , nan=SCREAMING_SNAKE_CASE__ ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { """predictions""": datasets.Sequence(datasets.Sequence(datasets.Value("""uint16""" ) ) ), """references""": datasets.Sequence(datasets.Sequence(datasets.Value("""uint16""" ) ) ), } ) ,reference_urls=[ """https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py""" ] ,) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ,snake_case ,snake_case ,snake_case = None ,snake_case = None ,snake_case = False ,): '''simple docstring''' lowercase : Union[str, Any] = mean_iou( results=snake_case ,gt_seg_maps=snake_case ,num_labels=snake_case ,ignore_index=snake_case ,nan_to_num=snake_case ,label_map=snake_case ,reduce_labels=snake_case ,) return iou_result
20
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
53
0
from math import factorial, pi def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ = 30 ) -> float: if not isinstance(lowerCamelCase_ , (int, float) ): raise ValueError('maclaurin_sin() requires either an int or float for theta' ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or accuracy <= 0: raise ValueError('maclaurin_sin() requires a positive int for accuracy' ) _lowercase : Dict = float(lowerCamelCase_ ) _lowercase : Union[str, Any] = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(lowerCamelCase_ ) ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ = 30 ) -> float: if not isinstance(lowerCamelCase_ , (int, float) ): raise ValueError('maclaurin_cos() requires either an int or float for theta' ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or accuracy <= 0: raise ValueError('maclaurin_cos() requires a positive int for accuracy' ) _lowercase : Union[str, Any] = float(lowerCamelCase_ ) _lowercase : Union[str, Any] = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(lowerCamelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(10)) print(maclaurin_sin(-10)) print(maclaurin_sin(10, 15)) print(maclaurin_sin(-10, 15)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(10, 15)) print(maclaurin_cos(-10, 15))
21
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowercase__ ( __lowercase : Features ) -> Optional[int]: """simple docstring""" __UpperCamelCase = np.inf def set_batch_size(__lowercase : FeatureType ) -> None: nonlocal batch_size if isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__lowercase , __lowercase ) and feature.dtype == "binary": __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__lowercase , __lowercase ) return None if batch_size is np.inf else batch_size class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , __A : NestedDataStructureLike[PathLike] , __A : Optional[NamedSplit] = None , __A : Optional[Features] = None , __A : str = None , __A : bool = False , __A : bool = False , __A : Optional[int] = None , **__A : Dict , ): super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) __UpperCamelCase = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} __UpperCamelCase = _PACKAGED_DATASETS_MODULES['parquet'][1] __UpperCamelCase = Parquet( cache_dir=__A , data_files=__A , features=__A , hash=__A , **__A , ) def _lowerCamelCase ( self : Optional[int] ): # Build iterable dataset if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Dataset , __A : Union[PathLike, BinaryIO] , __A : Optional[int] = None , **__A : Dict , ): __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size or get_writer_batch_size(dataset.features ) __UpperCamelCase = parquet_writer_kwargs def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , 'wb+' ) as buffer: __UpperCamelCase = self._write(file_obj=__A , batch_size=__A , **self.parquet_writer_kwargs ) else: __UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__A , **self.parquet_writer_kwargs ) return written def _lowerCamelCase ( self : List[str] , __A : BinaryIO , __A : int , **__A : List[str] ): __UpperCamelCase = 0 __UpperCamelCase = parquet_writer_kwargs.pop('path_or_buf' , __A ) __UpperCamelCase = self.dataset.features.arrow_schema __UpperCamelCase = pq.ParquetWriter(__A , schema=__A , **__A ) for offset in logging.tqdm( range(0 , len(self.dataset ) , __A ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating parquet from Arrow format' , ): __UpperCamelCase = query_table( table=self.dataset._data , key=slice(__A , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__A ) written += batch.nbytes writer.close() return written
53
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE :Tuple = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE :List[str] = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json''' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class A_ ( lowerCAmelCase_ ): _lowerCamelCase : int = """trocr""" _lowerCamelCase : List[str] = ["""past_key_values"""] _lowerCamelCase : int = { """num_attention_heads""": """decoder_attention_heads""", """hidden_size""": """d_model""", """num_hidden_layers""": """decoder_layers""", } def __init__( self : Union[str, Any] , snake_case_ : Dict=5_0_2_6_5 , snake_case_ : Dict=1_0_2_4 , snake_case_ : Optional[Any]=1_2 , snake_case_ : int=1_6 , snake_case_ : Tuple=4_0_9_6 , snake_case_ : List[Any]="gelu" , snake_case_ : Dict=5_1_2 , snake_case_ : str=0.1 , snake_case_ : Optional[Any]=0.0 , snake_case_ : Optional[int]=0.0 , snake_case_ : List[str]=2 , snake_case_ : int=0.0_2 , snake_case_ : int=0.0 , snake_case_ : Union[str, Any]=True , snake_case_ : str=False , snake_case_ : Tuple=True , snake_case_ : List[str]=True , snake_case_ : Optional[int]=1 , snake_case_ : int=0 , snake_case_ : Dict=2 , **snake_case_ : List[Any] , ): _UpperCAmelCase = vocab_size _UpperCAmelCase = d_model _UpperCAmelCase = decoder_layers _UpperCAmelCase = decoder_attention_heads _UpperCAmelCase = decoder_ffn_dim _UpperCAmelCase = activation_function _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation_dropout _UpperCAmelCase = init_std _UpperCAmelCase = decoder_layerdrop _UpperCAmelCase = use_cache _UpperCAmelCase = scale_embedding _UpperCAmelCase = use_learned_position_embeddings _UpperCAmelCase = layernorm_embedding super().__init__( pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , **snake_case_ , )
22
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def lowercase__ ( __lowercase : SplitDict ) -> int: """simple docstring""" __UpperCamelCase = split_dict._to_yaml_list() assert len(__lowercase ) == len(__lowercase ) __UpperCamelCase = SplitDict._from_yaml_list(__lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __UpperCamelCase = None # the split name of split_dict takes over the name of the split info object __UpperCamelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=__lowercase ), SplitInfo(dataset_name='my_dataset' )] ) def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" __UpperCamelCase = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
53
0
'''simple docstring''' def snake_case_ ( _lowerCAmelCase : list[int] ) -> float: if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) UpperCAmelCase : Tuple = sum(_lowerCAmelCase ) / len(_lowerCAmelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[str] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Any =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys a__ : str =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_available, logging from .benchmark_utils import ( Benchmark, Memory, MemorySummary, measure_peak_memory_cpu, start_memory_tracing, stop_memory_tracing, ) if is_tf_available(): import tensorflow as tf from tensorflow.python.framework.errors_impl import ResourceExhaustedError from .benchmark_args_tf import TensorFlowBenchmarkArguments if is_pyanvml_available(): import pyanvml.pyanvml as nvml snake_case_ = logging.get_logger(__name__) def lowerCamelCase__ ( snake_case_ : bool , snake_case_ : bool ) -> Optional[Any]: def run_func(snake_case_ : Union[str, Any] ): @wraps(snake_case_ ) def run_in_eager_mode(*snake_case_ : str , **snake_case_ : Any ): return func(*snake_case_ , **snake_case_ ) @wraps(snake_case_ ) @tf.function(experimental_compile=snake_case_ ) def run_in_graph_mode(*snake_case_ : List[str] , **snake_case_ : Any ): return func(*snake_case_ , **snake_case_ ) if do_eager_mode is True: if use_xla is not False: raise ValueError( '''Cannot run model in XLA, if `args.eager_mode` is set to `True`. Please set `args.eager_mode=False`.''' ) return run_in_eager_mode else: return run_in_graph_mode return run_func def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : int ) -> ["tf.Tensor"]: __snake_case = random.Random() __snake_case = [rng.randint(0 , vocab_size - 1 ) for i in range(batch_size * sequence_length )] return tf.constant(snake_case_ , shape=(batch_size, sequence_length) , dtype=tf.intaa ) class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): A_ : TensorFlowBenchmarkArguments A_ : PretrainedConfig A_ : str = "TensorFlow" @property def a (self : str ): """simple docstring""" return tf.__version__ def a (self : Optional[int] , a__ : str , a__ : int , a__ : int ): """simple docstring""" __snake_case = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) __snake_case = self._prepare_inference_func(a__ , a__ , a__ ) return self._measure_speed(_inference ) def a (self : Dict , a__ : str , a__ : int , a__ : int ): """simple docstring""" __snake_case = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) __snake_case = self._prepare_train_func(a__ , a__ , a__ ) return self._measure_speed(_train ) def a (self : List[str] , a__ : str , a__ : int , a__ : int ): """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , a__ ) __snake_case = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) __snake_case = self._prepare_inference_func(a__ , a__ , a__ ) return self._measure_memory(_inference ) def a (self : Tuple , a__ : str , a__ : int , a__ : int ): """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , a__ ) __snake_case = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) __snake_case = self._prepare_train_func(a__ , a__ , a__ ) return self._measure_memory(_train ) def a (self : Union[str, Any] , a__ : str , a__ : int , a__ : int ): """simple docstring""" __snake_case = self.config_dict[model_name] if self.args.fpaa: raise NotImplementedError('''Mixed precision is currently not supported.''' ) __snake_case = ( hasattr(a__ , '''architectures''' ) and isinstance(config.architectures , a__ ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: __snake_case = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model __snake_case = __import__('''transformers''' , fromlist=[model_class] ) __snake_case = getattr(a__ , a__ ) __snake_case = model_cls(a__ ) except ImportError: raise ImportError( f"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" ''' set `--only_pretrain_model` or `args.only_pretrain_model=True`.''' ) else: __snake_case = TF_MODEL_MAPPING[config.__class__](a__ ) # encoder-decoder has vocab size saved differently __snake_case = config.vocab_size if hasattr(a__ , '''vocab_size''' ) else config.encoder.vocab_size __snake_case = random_input_ids(a__ , a__ , a__ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_forward(): return model(a__ , decoder_input_ids=a__ , training=a__ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_forward(): return model(a__ , training=a__ ) __snake_case = encoder_decoder_forward if config.is_encoder_decoder else encoder_forward return _inference def a (self : Union[str, Any] , a__ : str , a__ : int , a__ : int ): """simple docstring""" __snake_case = self.config_dict[model_name] if self.args.eager_mode is not False: raise ValueError('''Training cannot be done in eager mode. Please make sure that `args.eager_mode = False`.''' ) if self.args.fpaa: raise NotImplementedError('''Mixed precision is currently not supported.''' ) __snake_case = ( hasattr(a__ , '''architectures''' ) and isinstance(config.architectures , a__ ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: __snake_case = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model __snake_case = __import__('''transformers''' , fromlist=[model_class] ) __snake_case = getattr(a__ , a__ ) __snake_case = model_cls(a__ ) except ImportError: raise ImportError( f"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" ''' set `--only_pretrain_model` or `args.only_pretrain_model=True`.''' ) else: __snake_case = TF_MODEL_WITH_LM_HEAD_MAPPING[config.__class__](a__ ) # encoder-decoder has vocab size saved differently __snake_case = config.vocab_size if hasattr(a__ , '''vocab_size''' ) else config.encoder.vocab_size __snake_case = random_input_ids(a__ , a__ , a__ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_train(): __snake_case = model(a__ , decoder_input_ids=a__ , labels=a__ , training=a__ )[0] __snake_case = tf.gradients(a__ , model.trainable_variables ) return gradients @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_train(): __snake_case = model(a__ , labels=a__ , training=a__ )[0] __snake_case = tf.gradients(a__ , model.trainable_variables ) return gradients __snake_case = encoder_decoder_train if config.is_encoder_decoder else encoder_train return _train def a (self : List[Any] , a__ : Dict ): """simple docstring""" with self.args.strategy.scope(): try: if self.args.is_tpu or self.args.use_xla: # run additional 10 times to stabilize compilation for tpu logger.info('''Do inference on TPU. Running model 5 times to stabilize compilation''' ) timeit.repeat(a__ , repeat=1 , number=5 ) # as written in https://docs.python.org/2/library/timeit.html#timeit.Timer.repeat, min should be taken rather than the average __snake_case = timeit.repeat( a__ , repeat=self.args.repeat , number=10 , ) return min(a__ ) / 1_0.0 except ResourceExhaustedError as e: self.print_fn(f"""Doesn't fit on GPU. {e}""" ) def a (self : Dict , a__ : Callable[[], None] ): """simple docstring""" logger.info( '''Note that TensorFlow allocates more memory than ''' '''it might need to speed up computation. ''' '''The memory reported here corresponds to the memory ''' '''reported by `nvidia-smi`, which can vary depending ''' '''on total available memory on the GPU that is used.''' ) with self.args.strategy.scope(): try: if self.args.trace_memory_line_by_line: if not self.args.eager_mode: raise ValueError( '''`args.eager_mode` is set to `False`. Make sure to run model in eager mode to measure memory''' ''' consumption line by line.''' ) __snake_case = start_memory_tracing('''transformers''' ) if self.args.is_tpu: # tpu raise NotImplementedError( '''Memory Benchmarking is currently not implemented for TPU. Please disable memory benchmarking''' ''' with `args.memory=False`''' ) elif self.args.is_gpu: # gpu if not is_pyanvml_available(): logger.warning( '''py3nvml not installed, we won\'t log GPU memory usage. ''' '''Install py3nvml (pip install py3nvml) to log information about GPU.''' ) __snake_case = '''N/A''' else: logger.info( '''Measuring total GPU usage on GPU device. Make sure to not have additional processes''' ''' running on the same GPU.''' ) # init nvml nvml.nvmlInit() func() __snake_case = nvml.nvmlDeviceGetHandleByIndex(self.args.device_idx ) __snake_case = nvml.nvmlDeviceGetMemoryInfo(a__ ) __snake_case = meminfo.used __snake_case = Memory(a__ ) # shutdown nvml nvml.nvmlShutdown() else: # cpu if self.args.trace_memory_line_by_line: logger.info( '''When enabling line by line tracing, the max peak memory for CPU is inaccurate in''' ''' TensorFlow.''' ) __snake_case = None else: __snake_case = measure_peak_memory_cpu(a__ ) __snake_case = Memory(a__ ) if isinstance(a__ , a__ ) else memory_bytes if self.args.trace_memory_line_by_line: __snake_case = stop_memory_tracing(a__ ) if memory is None: __snake_case = summary.total else: __snake_case = None return memory, summary except ResourceExhaustedError as e: self.print_fn(f"""Doesn't fit on GPU. {e}""" ) return "N/A", None
24
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a__ : str =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =["input_features", "attention_mask"] def __init__( self : Union[str, Any] , __A : Optional[int]=8_0 , __A : Tuple=1_6_0_0_0 , __A : Optional[Any]=8_0 , __A : Any=0.0 , __A : Any=True , __A : List[str]=True , __A : str=True , **__A : List[Any] , ): super().__init__(feature_size=__A , sampling_rate=__A , padding_value=__A , **__A ) __UpperCamelCase = num_mel_bins __UpperCamelCase = do_ceptral_normalize __UpperCamelCase = normalize_means __UpperCamelCase = normalize_vars __UpperCamelCase = True def _lowerCamelCase ( self : Union[str, Any] , __A : np.ndarray , ): __UpperCamelCase = waveform * (2**1_5) # Kaldi compliance: 16-bit signed integers __UpperCamelCase = torch.from_numpy(__A ).unsqueeze(0 ) __UpperCamelCase = ta_kaldi.fbank(__A , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( __A : np.ndarray , __A : int , __A : Optional[bool] = True , __A : Optional[bool] = True , __A : float = 0.0 , ): # make sure we normalize float32 arrays if normalize_means: __UpperCamelCase = x[:input_length].mean(axis=0 ) __UpperCamelCase = np.subtract(__A , __A ) if normalize_vars: __UpperCamelCase = x[:input_length].std(axis=0 ) __UpperCamelCase = np.divide(__A , __A ) if input_length < x.shape[0]: __UpperCamelCase = padding_value # make sure array is in float32 __UpperCamelCase = x.astype(np.floataa ) return x def _lowerCamelCase ( self : int , __A : List[np.ndarray] , __A : Optional[np.ndarray] = None ): __UpperCamelCase = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(__A , __A , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(__A , __A ) ] def __call__( self : List[Any] , __A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __A : Union[bool, str, PaddingStrategy] = False , __A : Optional[int] = None , __A : bool = False , __A : Optional[int] = None , __A : Optional[Union[str, TensorType]] = None , __A : Optional[int] = None , __A : Optional[bool] = None , **__A : Dict , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) __UpperCamelCase = isinstance(__A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __UpperCamelCase = is_batched_numpy or ( isinstance(__A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(__A , np.ndarray ): __UpperCamelCase = np.asarray(__A , dtype=np.floataa ) elif isinstance(__A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCamelCase = [raw_speech] # extract fbank features __UpperCamelCase = [self._extract_fbank_features(__A ) for waveform in raw_speech] # convert into correct format for padding __UpperCamelCase = BatchFeature({'input_features': features} ) __UpperCamelCase = self.pad( __A , padding=__A , max_length=__A , truncation=__A , pad_to_multiple_of=__A , return_attention_mask=__A , **__A , ) # make sure list is in array format __UpperCamelCase = padded_inputs.get('input_features' ) if isinstance(input_features[0] , __A ): __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for feature in input_features] __UpperCamelCase = padded_inputs.get('attention_mask' ) if attention_mask is not None: __UpperCamelCase = [np.asarray(__A , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __UpperCamelCase = ( np.array(__A , dtype=np.intaa ) if self._get_padding_strategies(__A , max_length=__A ) is not PaddingStrategy.DO_NOT_PAD else None ) __UpperCamelCase = self.normalize( padded_inputs['input_features'] , attention_mask=__A ) if return_tensors is not None: __UpperCamelCase = padded_inputs.convert_to_tensors(__A ) return padded_inputs
53
0
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class lowerCAmelCase_ (a__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Optional[Any] = MvpTokenizer __UpperCamelCase : Any = MvpTokenizerFast __UpperCamelCase : Optional[int] = True __UpperCamelCase : int = filter_roberta_detectors def __magic_name__ (self ) -> int: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ : Optional[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] SCREAMING_SNAKE_CASE__ : List[str] = dict(zip(SCREAMING_SNAKE_CASE__ , range(len(SCREAMING_SNAKE_CASE__ ) ) ) ) SCREAMING_SNAKE_CASE__ : int = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] SCREAMING_SNAKE_CASE__ : Optional[int] = {"""unk_token""": """<unk>"""} SCREAMING_SNAKE_CASE__ : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE__ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(SCREAMING_SNAKE_CASE__ ) ) def __magic_name__ (self , **SCREAMING_SNAKE_CASE__ ) -> List[Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , **SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> int: """simple docstring""" return "lower newer", "lower newer" @cached_property def __magic_name__ (self ) -> Any: """simple docstring""" return MvpTokenizer.from_pretrained("""RUCAIBox/mvp""" ) @cached_property def __magic_name__ (self ) -> List[str]: """simple docstring""" return MvpTokenizerFast.from_pretrained("""RUCAIBox/mvp""" ) @require_torch def __magic_name__ (self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] SCREAMING_SNAKE_CASE__ : List[str] = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE__ : Any = tokenizer(SCREAMING_SNAKE_CASE__ , max_length=len(SCREAMING_SNAKE_CASE__ ) , padding=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) SCREAMING_SNAKE_CASE__ : List[Any] = batch.input_ids.tolist()[0] self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Test that special tokens are reset @require_torch def __magic_name__ (self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer(SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) # check if input_ids are returned and no labels self.assertIn("""input_ids""" , SCREAMING_SNAKE_CASE__ ) self.assertIn("""attention_mask""" , SCREAMING_SNAKE_CASE__ ) self.assertNotIn("""labels""" , SCREAMING_SNAKE_CASE__ ) self.assertNotIn("""decoder_attention_mask""" , SCREAMING_SNAKE_CASE__ ) @require_torch def __magic_name__ (self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE__ : Dict = tokenizer(text_target=SCREAMING_SNAKE_CASE__ , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def __magic_name__ (self ) -> Dict: """simple docstring""" for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer( ["""I am a small frog""" * 10_24, """I am a small frog"""] , padding=SCREAMING_SNAKE_CASE__ , truncation=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) self.assertIsInstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) self.assertEqual(batch.input_ids.shape , (2, 10_24) ) @require_torch def __magic_name__ (self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = ["""A long paragraph for summarization."""] SCREAMING_SNAKE_CASE__ : Optional[Any] = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer(SCREAMING_SNAKE_CASE__ , text_target=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ : List[Any] = inputs["""input_ids"""] SCREAMING_SNAKE_CASE__ : Optional[Any] = inputs["""labels"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def __magic_name__ (self ) -> Dict: """simple docstring""" pass def __magic_name__ (self ) -> Tuple: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = self.tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = """A, <mask> AllenNLP sentence.""" SCREAMING_SNAKE_CASE__ : int = tokenizer_r.encode_plus(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ , return_token_type_ids=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = tokenizer_p.encode_plus(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ , return_token_type_ids=SCREAMING_SNAKE_CASE__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( SCREAMING_SNAKE_CASE__ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( SCREAMING_SNAKE_CASE__ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
25
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : List[Any] =logging.get_logger(__name__) a__ : List[Any] ={ '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_text_model" def __init__( self : str , __A : List[Any]=2_5_0_0_0_2 , __A : Any=1_0_2_4 , __A : int=2_4 , __A : Dict=1_6 , __A : Optional[Any]=4_0_9_6 , __A : Union[str, Any]="gelu" , __A : Dict=0.1 , __A : Dict=0.1 , __A : List[str]=5_1_4 , __A : Optional[int]=1 , __A : int=0.02 , __A : Optional[Any]=0.02 , __A : Optional[Any]=1e-05 , __A : Dict=1 , __A : List[Any]=0 , __A : int=2 , __A : Tuple="absolute" , __A : Optional[Any]=True , __A : Optional[int]=7_6_8 , **__A : List[str] , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = use_cache __UpperCamelCase = project_dim class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_vision_model" def __init__( self : List[Any] , __A : Union[str, Any]=7_6_8 , __A : Optional[int]=3_0_7_2 , __A : Optional[Any]=5_1_2 , __A : Tuple=1_2 , __A : Union[str, Any]=1_2 , __A : Optional[int]=3 , __A : Dict=2_2_4 , __A : Tuple=3_2 , __A : str="quick_gelu" , __A : Dict=1e-5 , __A : Optional[int]=0.0 , __A : List[Any]=0.02 , __A : int=1.0 , **__A : Optional[int] , ): super().__init__(**__A ) __UpperCamelCase = hidden_size __UpperCamelCase = intermediate_size __UpperCamelCase = projection_dim __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = num_channels __UpperCamelCase = patch_size __UpperCamelCase = image_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = attention_dropout __UpperCamelCase = layer_norm_eps __UpperCamelCase = hidden_act @classmethod def _lowerCamelCase ( cls : Optional[Any] , __A : Union[str, os.PathLike] , **__A : Optional[Any] ): cls._set_token_in_kwargs(__A ) __UpperCamelCase , __UpperCamelCase = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('model_type' ) == "altclip": __UpperCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] ="altclip" SCREAMING_SNAKE_CASE_ : Optional[int] =True def __init__( self : Any , __A : List[str]=None , __A : List[Any]=None , __A : List[str]=7_6_8 , __A : List[str]=2.6592 , **__A : Dict ): # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __UpperCamelCase = kwargs.pop('text_config_dict' , __A ) __UpperCamelCase = kwargs.pop('vision_config_dict' , __A ) super().__init__(**__A ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __UpperCamelCase = {} # This is the complete result when using `text_config_dict`. __UpperCamelCase = AltCLIPTextConfig(**__A ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `text_config_dict` and `text_config` but with different values. ''' f'''The value `text_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The ''' f'''value `text_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __UpperCamelCase = {} # This is the complete result when using `vision_config_dict`. __UpperCamelCase = AltCLIPVisionConfig(**__A ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __UpperCamelCase = { str(__A ): value for key, value in _vision_config_dict['id2label'].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `vision_config_dict` and `vision_config` but with different ''' f'''values. The value `vision_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. ''' f'''The value `vision_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __UpperCamelCase = {} logger.info('`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.' ) if vision_config is None: __UpperCamelCase = {} logger.info('`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.' ) __UpperCamelCase = AltCLIPTextConfig(**__A ) __UpperCamelCase = AltCLIPVisionConfig(**__A ) __UpperCamelCase = projection_dim __UpperCamelCase = logit_scale_init_value __UpperCamelCase = 1.0 @classmethod def _lowerCamelCase ( cls : Union[str, Any] , __A : AltCLIPTextConfig , __A : AltCLIPVisionConfig , **__A : Optional[Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) __UpperCamelCase = self.text_config.to_dict() __UpperCamelCase = self.vision_config.to_dict() __UpperCamelCase = self.__class__.model_type return output
53
0
from __future__ import annotations from typing import Generic, TypeVar _snake_case = TypeVar("T") class lowercase ( Generic[T] ): def __init__( self , _a ) -> None: _A : List[str] = data _A : int = self _A : Any = 0 class lowercase ( Generic[T] ): def __init__( self ) -> None: # map from node name to the node object _A : dict[T, DisjointSetTreeNode[T]] = {} def a__ ( self , _a ) -> None: # create a new set with x as its member _A : Union[str, Any] = DisjointSetTreeNode(_a ) def a__ ( self , _a ) -> DisjointSetTreeNode[T]: # find the set x belongs to (with path-compression) _A : Any = self.map[data] if elem_ref != elem_ref.parent: _A : Optional[int] = self.find_set(elem_ref.parent.data ) return elem_ref.parent def a__ ( self , _a , _a ) -> None: # helper function for union operation if nodea.rank > nodea.rank: _A : List[str] = nodea else: _A : str = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def a__ ( self , _a , _a ) -> None: # merge 2 disjoint sets self.link(self.find_set(_a ) , self.find_set(_a ) ) class lowercase ( Generic[T] ): def __init__( self ) -> None: # connections: map from the node to the neighbouring nodes (with weights) _A : dict[T, dict[T, int]] = {} def a__ ( self , _a ) -> None: # add a node ONLY if its not present in the graph if node not in self.connections: _A : Dict = {} def a__ ( self , _a , _a , _a ) -> None: # add an edge with the given weight self.add_node(_a ) self.add_node(_a ) _A : Optional[int] = weight _A : Tuple = weight def a__ ( self ) -> GraphUndirectedWeighted[T]: _A : Dict = [] _A : Dict = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda _a : x[2] ) # creating the disjoint set _A : Any = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(_a ) # MST generation _A : Optional[int] = 0 _A : List[Any] = 0 _A : Dict = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: _A , _A , _A : Tuple = edges[index] index += 1 _A : Optional[Any] = disjoint_set.find_set(_a ) _A : List[str] = disjoint_set.find_set(_a ) if parent_u != parent_v: num_edges += 1 graph.add_edge(_a , _a , _a ) disjoint_set.union(_a , _a ) return graph
26
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Any ) -> Optional[Any]: """simple docstring""" with open(__lowercase ) as metadata_file: __UpperCamelCase = json.load(__lowercase ) __UpperCamelCase = LukeConfig(use_entity_aware_attention=__lowercase , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __UpperCamelCase = torch.load(__lowercase , map_location='cpu' ) # Load the entity vocab file __UpperCamelCase = load_entity_vocab(__lowercase ) __UpperCamelCase = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __UpperCamelCase = AddedToken('<ent>' , lstrip=__lowercase , rstrip=__lowercase ) __UpperCamelCase = AddedToken('<ent2>' , lstrip=__lowercase , rstrip=__lowercase ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(__lowercase , __lowercase ) __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase ) # Initialize the embeddings of the special tokens __UpperCamelCase = state_dict['embeddings.word_embeddings.weight'] __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __UpperCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __UpperCamelCase = F'''encoder.layer.{layer_index}.attention.self.''' __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __UpperCamelCase = state_dict['entity_embeddings.entity_embeddings.weight'] __UpperCamelCase = entity_emb[entity_vocab['[MASK]']] __UpperCamelCase = LukeModel(config=__lowercase ).eval() __UpperCamelCase , __UpperCamelCase = model.load_state_dict(__lowercase , strict=__lowercase ) if not (len(__lowercase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {', '.join(__lowercase )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' F''' {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}''' ) # Check outputs __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase , task='entity_classification' ) __UpperCamelCase = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __UpperCamelCase = (39, 42) __UpperCamelCase = tokenizer(__lowercase , entity_spans=[span] , add_prefix_space=__lowercase , return_tensors='pt' ) __UpperCamelCase = model(**__lowercase ) # Verify word hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 42, 1024) ) __UpperCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base __UpperCamelCase = torch.Size((1, 42, 768) ) __UpperCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 1, 1024) ) __UpperCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base __UpperCamelCase = torch.Size((1, 1, 768) ) __UpperCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(__lowercase ) ) model.save_pretrained(__lowercase ) def lowercase__ ( __lowercase : Dict ) -> List[str]: """simple docstring""" __UpperCamelCase = {} with open(__lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(__lowercase ): __UpperCamelCase , __UpperCamelCase = line.rstrip().split('\t' ) __UpperCamelCase = index return entity_vocab if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : str =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
53
0
'''simple docstring''' from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class __UpperCamelCase ( lowerCAmelCase_ ): A_ = "EncodecFeatureExtractor" A_ = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , __a , __a ): '''simple docstring''' super().__init__(__a , __a ) __a : Tuple = self.feature_extractor __a : Dict = False def __UpperCAmelCase ( self , __a=None , __a=None , __a=True ): '''simple docstring''' return self.tokenizer.get_decoder_prompt_ids(task=__a , language=__a , no_timestamps=__a ) def __call__( self , *__a , **__a ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*__a , **__a ) __a : Optional[Any] = kwargs.pop('audio' , __a ) __a : Union[str, Any] = kwargs.pop('sampling_rate' , __a ) __a : Optional[Any] = kwargs.pop('text' , __a ) if len(__a ) > 0: __a : Any = args[0] __a : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if text is not None: __a : int = self.tokenizer(__a , **__a ) if audio is not None: __a : Optional[int] = self.feature_extractor(__a , *__a , sampling_rate=__a , **__a ) if audio is None: return inputs elif text is None: return audio_inputs else: __a : Dict = audio_inputs['input_values'] if "padding_mask" in audio_inputs: __a : Optional[Any] = audio_inputs['padding_mask'] return inputs def __UpperCAmelCase ( self , *__a , **__a ): '''simple docstring''' __a : Tuple = kwargs.pop('audio' , __a ) __a : int = kwargs.pop('padding_mask' , __a ) if len(__a ) > 0: __a : Any = args[0] __a : List[str] = args[1:] if audio_values is not None: return self._decode_audio(__a , padding_mask=__a ) else: return self.tokenizer.batch_decode(*__a , **__a ) def __UpperCAmelCase ( self , *__a , **__a ): '''simple docstring''' return self.tokenizer.decode(*__a , **__a ) def __UpperCAmelCase ( self , __a , __a = None ): '''simple docstring''' __a : str = to_numpy(__a ) __a , __a , __a : Union[str, Any] = audio_values.shape if padding_mask is None: return list(__a ) __a : Optional[int] = to_numpy(__a ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __a : Union[str, Any] = seq_len - padding_mask.shape[-1] __a : List[str] = 1 - self.feature_extractor.padding_value __a : List[Any] = np.pad(__a , ((0, 0), (0, difference)) , 'constant' , constant_values=__a ) __a : Optional[int] = audio_values.tolist() for i in range(__a ): __a : Any = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __a : Optional[Any] = sliced_audio.reshape(__a , -1 ) return audio_values
27
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case ( __lowerCamelCase ): """simple docstring""" def _lowerCamelCase ( self : Any ): __UpperCamelCase = tempfile.mkdtemp() __UpperCamelCase = 8 # DPR tok __UpperCamelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Optional[int] ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Union[str, Any] ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def _lowerCamelCase ( self : str ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _lowerCamelCase ( self : Any , __A : bool ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __UpperCamelCase = os.path.join(self.tmpdirname , 'dataset' ) __UpperCamelCase = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _lowerCamelCase ( self : int ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __UpperCamelCase = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __UpperCamelCase = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __UpperCamelCase = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__A , open(__A , 'wb' ) ) __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_legacy_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __A ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Optional[Any] ): import torch __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) __UpperCamelCase = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors='pt' , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dpr_ctx_encoder_tokenizer() __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __A ) # check for doc token related keys in dictionary.
53
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCamelCase : Tuple = { "configuration_blip": [ "BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlipConfig", "BlipTextConfig", "BlipVisionConfig", ], "processing_blip": ["BlipProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Any = ["BlipImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = [ "BLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "BlipModel", "BlipPreTrainedModel", "BlipForConditionalGeneration", "BlipForQuestionAnswering", "BlipVisionModel", "BlipTextModel", "BlipForImageTextRetrieval", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Union[str, Any] = [ "TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "TFBlipModel", "TFBlipPreTrainedModel", "TFBlipForConditionalGeneration", "TFBlipForQuestionAnswering", "TFBlipVisionModel", "TFBlipTextModel", "TFBlipForImageTextRetrieval", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
28
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[Any] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys a__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup __UpperCAmelCase = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582' } def lowercase__ ( __snake_case : str = "dhaka" , __snake_case : int = 5 ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = min(__snake_case , 50 ) # Prevent abuse! UpperCAmelCase_ : Dict = { 'q': query, 'tbm': 'isch', 'hl': 'en', 'ijn': '0', } UpperCAmelCase_ : Any = requests.get('https://www.google.com/search' , params=__snake_case , headers=__snake_case ) UpperCAmelCase_ : Dict = BeautifulSoup(html.text , 'html.parser' ) UpperCAmelCase_ : Any = ''.join( re.findall(R'AF_initDataCallback\(([^<]+)\);' , str(soup.select('script' ) ) ) ) UpperCAmelCase_ : int = json.dumps(__snake_case ) UpperCAmelCase_ : List[Any] = json.loads(__snake_case ) UpperCAmelCase_ : Union[str, Any] = re.findall( R'\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\",' , __snake_case , ) if not matched_google_image_data: return 0 UpperCAmelCase_ : Union[str, Any] = re.sub( R'\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]' , '' , str(__snake_case ) , ) UpperCAmelCase_ : Optional[int] = re.findall( R'(?:\'|,),\[\"(https:|http.*?)\",\d+,\d+\]' , __snake_case , ) for index, fixed_full_res_image in enumerate(__snake_case ): if index >= max_images: return index UpperCAmelCase_ : Optional[int] = bytes(__snake_case , 'ascii' ).decode( 'unicode-escape' ) UpperCAmelCase_ : Union[str, Any] = bytes(__snake_case , 'ascii' ).decode( 'unicode-escape' ) UpperCAmelCase_ : Union[str, Any] = urllib.request.build_opener() UpperCAmelCase_ : Dict = [ ( 'User-Agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582', ) ] urllib.request.install_opener(__snake_case ) UpperCAmelCase_ : Union[str, Any] = F"query_{query.replace(' ' , '_' )}" if not os.path.exists(__snake_case ): os.makedirs(__snake_case ) urllib.request.urlretrieve( # noqa: S310 __snake_case , F"{path_name}/original_size_img_{index}.jpg" ) return index if __name__ == "__main__": try: __UpperCAmelCase = download_images_from_google_query(sys.argv[1]) print(F'{image_count} images were downloaded to disk.') except IndexError: print('Please provide a search term.') raise
29
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
0
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness __a = '\\n@misc{chen2021evaluating,\n title={Evaluating Large Language Models Trained on Code},\n author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \\nand Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \\nand Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \\nand Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \\nand Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \\nand Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \\nand Mohammad Bavarian and Clemens Winter and Philippe Tillet \\nand Felipe Petroski Such and Dave Cummings and Matthias Plappert \\nand Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \\nand William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \\nand Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \\nand William Saunders and Christopher Hesse and Andrew N. Carr \\nand Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \\nand Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \\nand Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \\nand Sam McCandlish and Ilya Sutskever and Wojciech Zaremba},\n year={2021},\n eprint={2107.03374},\n archivePrefix={arXiv},\n primaryClass={cs.LG}\n}\n' __a = '\\nThis metric implements the evaluation harness for the HumanEval problem solving dataset\ndescribed in the paper "Evaluating Large Language Models Trained on Code"\n(https://arxiv.org/abs/2107.03374).\n' __a = '\nCalculates how good are predictions given some references, using certain scores\nArgs:\n predictions: list of candidates to evaluate. Each candidates should be a list\n of strings with several code candidates to solve the problem.\n references: a list with a test for each prediction. Each test should evaluate the\n correctness of a code candidate.\n k: number of code candidates to consider in the evaluation (Default: [1, 10, 100])\n num_workers: number of workers used to evaluate the canidate programs (Default: 4).\n timeout:\nReturns:\n pass_at_k: dict with pass rates for each k\n results: dict with granular results of each unittest\nExamples:\n >>> code_eval = datasets.load_metric("code_eval")\n >>> test_cases = ["assert add(2,3)==5"]\n >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]]\n >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2])\n >>> print(pass_at_k)\n {\'pass@1\': 0.5, \'pass@2\': 1.0}\n' __a = '\n################################################################################\n !!!WARNING!!!\n################################################################################\nThe "code_eval" metric executes untrusted model-generated code in Python.\nAlthough it is highly unlikely that model-generated code will do something\novertly malicious in response to this test suite, model-generated code may act\ndestructively due to a lack of model capability or alignment.\nUsers are strongly encouraged to sandbox this evaluation suite so that it\ndoes not perform destructive actions on their host or network. For more\ninformation on how OpenAI sandboxes its code, see the paper "Evaluating Large\nLanguage Models Trained on Code" (https://arxiv.org/abs/2107.03374).\n\nOnce you have read this disclaimer and taken appropriate precautions,\nset the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this\nwith:\n\n>>> import os\n>>> os.environ["HF_ALLOW_CODE_EVAL"] = "1"\n\n################################################################################\\n' __a = 'The MIT License\n\nCopyright (c) OpenAI (https://openai.com)\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the "Software"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in\nall copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\nTHE SOFTWARE.' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase__( datasets.Metric ): """simple docstring""" def _lowercase ( self : int ) -> Dict: return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Value('''string''' ), } ) , homepage='''https://github.com/openai/human-eval''' , codebase_urls=['''https://github.com/openai/human-eval'''] , reference_urls=['''https://github.com/openai/human-eval'''] , license=_LICENSE , ) def _lowercase ( self : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Optional[int]=[1, 1_0, 1_0_0] , SCREAMING_SNAKE_CASE_ : str=4 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=3.0 ) -> int: if os.getenv('''HF_ALLOW_CODE_EVAL''' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('''This metric is currently not supported on Windows.''' ) with ThreadPoolExecutor(max_workers=SCREAMING_SNAKE_CASE_ ) as executor: lowercase_ = [] lowercase_ = Counter() lowercase_ = 0 lowercase_ = defaultdict(SCREAMING_SNAKE_CASE_ ) for task_id, (candidates, test_case) in enumerate(zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ): for candidate in candidates: lowercase_ = candidate + '''\n''' + test_case lowercase_ = (test_program, timeout, task_id, completion_id[task_id]) lowercase_ = executor.submit(SCREAMING_SNAKE_CASE_ , *SCREAMING_SNAKE_CASE_ ) futures.append(SCREAMING_SNAKE_CASE_ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(SCREAMING_SNAKE_CASE_ ): lowercase_ = future.result() results[result["task_id"]].append((result['''completion_id'''], result) ) lowercase_ , lowercase_ = [], [] for result in results.values(): result.sort() lowercase_ = [r[1]['''passed'''] for r in result] total.append(len(SCREAMING_SNAKE_CASE_ ) ) correct.append(sum(SCREAMING_SNAKE_CASE_ ) ) lowercase_ = np.array(SCREAMING_SNAKE_CASE_ ) lowercase_ = np.array(SCREAMING_SNAKE_CASE_ ) lowercase_ = k lowercase_ = {f'''pass@{k}''': estimate_pass_at_k(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def a ( snake_case__: Dict , snake_case__: List[str] , snake_case__: Union[str, Any] ): '''simple docstring''' def estimator(snake_case__: int , snake_case__: int , snake_case__: int ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(snake_case__ , snake_case__ ): lowercase_ = itertools.repeat(snake_case__ , len(snake_case__ ) ) else: assert len(snake_case__ ) == len(snake_case__ ) lowercase_ = iter(snake_case__ ) return np.array([estimator(int(snake_case__ ) , int(snake_case__ ) , snake_case__ ) for n, c in zip(snake_case__ , snake_case__ )] )
30
'''simple docstring''' # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys a__ : Tuple ='''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
53
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_ (snake_case__ ): '''simple docstring''' def __init__( self : Any , *A : List[str] , **A : Dict ): warnings.warn( "The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use BeitImageProcessor instead." , A , ) super().__init__(*A , **A )
31
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
0
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration UpperCAmelCase_ : str = HfArgumentParser(InitializationArguments) UpperCAmelCase_ : str = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization UpperCAmelCase_ : str = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks UpperCAmelCase_ : Optional[Any] = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) UpperCAmelCase_ : Dict = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config UpperCAmelCase_ : Optional[int] = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
32
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : List[Any] ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] =("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , __A : Optional[int] , __A : List[Any] ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : List[Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : List[Any] , *__A : Dict , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : List[Any] , *__A : List[str] , **__A : Dict ): return self.tokenizer.decode(*__A , **__A ) @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
0
"""simple docstring""" def lowercase ( __snake_case : int , __snake_case : int ): return 1 if input_a == input_a else 0 def lowercase ( ): assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
33
'''simple docstring''' from __future__ import annotations from typing import Any class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case : """simple docstring""" def __init__( self : List[Any] , __A : Any ): __UpperCamelCase = data __UpperCamelCase = None def __iter__( self : Optional[Any] ): __UpperCamelCase = self __UpperCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(__A ) yield node.data __UpperCamelCase = node.next_node @property def _lowerCamelCase ( self : List[str] ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": a__ : Dict =Node(1) a__ : Optional[int] =Node(2) a__ : List[str] =Node(3) a__ : Optional[int] =Node(4) print(root_node.has_loop) # False a__ : str =root_node.next_node print(root_node.has_loop) # True a__ : Optional[int] =Node(5) a__ : List[Any] =Node(6) a__ : int =Node(5) a__ : Tuple =Node(6) print(root_node.has_loop) # False a__ : str =Node(1) print(root_node.has_loop) # False
53
0
'''simple docstring''' import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _a ( __a ): def __init__( self : Optional[int] , *lowercase : Optional[Any] , lowercase : Optional[Any]=None , lowercase : Tuple=None , **lowercase : Dict ): '''simple docstring''' super().__init__(*lowercase , **lowercase ) UpperCAmelCase = eval_examples UpperCAmelCase = post_process_function def A ( self : str , lowercase : List[Any]=None , lowercase : Tuple=None , lowercase : Optional[Any]=None , lowercase : str = "eval" ): '''simple docstring''' UpperCAmelCase = self.eval_dataset if eval_dataset is None else eval_dataset UpperCAmelCase = self.get_eval_dataloader(lowercase ) UpperCAmelCase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop UpperCAmelCase = time.time() try: UpperCAmelCase = eval_loop( lowercase , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: UpperCAmelCase = compute_metrics UpperCAmelCase = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default UpperCAmelCase = self.post_process_function(lowercase , lowercase , output.predictions ) UpperCAmelCase = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): UpperCAmelCase = metrics.pop(lowercase ) metrics.update(output.metrics ) else: UpperCAmelCase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowercase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) UpperCAmelCase = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowercase ) return metrics def A ( self : Optional[Any] , lowercase : Optional[Any] , lowercase : Tuple , lowercase : Optional[Any]=None , lowercase : str = "test" ): '''simple docstring''' UpperCAmelCase = self.get_test_dataloader(lowercase ) # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop UpperCAmelCase = time.time() try: UpperCAmelCase = eval_loop( lowercase , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: UpperCAmelCase = compute_metrics UpperCAmelCase = self.args.eval_batch_size * self.args.world_size if f"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[f"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output UpperCAmelCase = self.post_process_function(lowercase , lowercase , output.predictions , '''predict''' ) UpperCAmelCase = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"{metric_key_prefix}_" ): UpperCAmelCase = metrics.pop(lowercase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowercase )
34
'''simple docstring''' a__ : Optional[Any] =256 # Modulus to hash a string a__ : Dict =1_000_003 def lowercase__ ( __lowercase : str , __lowercase : str ) -> bool: """simple docstring""" __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) if p_len > t_len: return False __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 1 # Calculating the hash of pattern and substring of text for i in range(__lowercase ): __UpperCamelCase = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus __UpperCamelCase = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue __UpperCamelCase = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash __UpperCamelCase = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: """simple docstring""" __UpperCamelCase = 'abc1abc12' __UpperCamelCase = 'alskfjaldsabc1abc1abc12k23adsfabcabc' __UpperCamelCase = 'alskfjaldsk23adsfabcabc' assert rabin_karp(__lowercase , __lowercase ) and not rabin_karp(__lowercase , __lowercase ) # Test 2) __UpperCamelCase = 'ABABX' __UpperCamelCase = 'ABABZABABYABABX' assert rabin_karp(__lowercase , __lowercase ) # Test 3) __UpperCamelCase = 'AAAB' __UpperCamelCase = 'ABAAAAAB' assert rabin_karp(__lowercase , __lowercase ) # Test 4) __UpperCamelCase = 'abcdabcy' __UpperCamelCase = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(__lowercase , __lowercase ) # Test 5) __UpperCamelCase = 'Lü' __UpperCamelCase = 'Lüsai' assert rabin_karp(__lowercase , __lowercase ) __UpperCamelCase = 'Lue' assert not rabin_karp(__lowercase , __lowercase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __a = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure)
35
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
0
from __future__ import annotations import requests _snake_case = set( "approved_at_utc approved_by author_flair_background_color\nauthor_flair_css_class author_flair_richtext author_flair_template_id author_fullname\nauthor_premium can_mod_post category clicked content_categories created_utc downs\nedited gilded gildings hidden hide_score is_created_from_ads_ui is_meta\nis_original_content is_reddit_media_domain is_video link_flair_css_class\nlink_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title\nname permalink pwls quarantine saved score secure_media secure_media_embed selftext\nsubreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type\ntotal_awards_received ups upvote_ratio url user_reports".split() ) def A ( _lowerCamelCase , _lowerCamelCase = 1 , _lowerCamelCase = "new" , _lowerCamelCase = None ): '''simple docstring''' _lowerCAmelCase : Optional[int] = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(_lowerCamelCase ) - valid_terms ) ): _lowerCAmelCase : int = F"Invalid search term: {invalid_search_terms}" raise ValueError(_lowerCamelCase ) _lowerCAmelCase : Optional[int] = requests.get( F"https://reddit.com/r/{subreddit}/{age}.json?limit={limit}" , headers={"User-agent": "A random string"} , ) if response.status_code == 429: raise requests.HTTPError _lowerCAmelCase : List[str] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(_lowerCamelCase )} _lowerCAmelCase : str = {} for id_ in range(_lowerCamelCase ): _lowerCAmelCase : Optional[int] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("learnpython", wanted_data=["title", "url", "selftext"]))
36
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
0
'''simple docstring''' from collections.abc import Callable def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : float = a lowerCAmelCase__ : float = b if function(UpperCamelCase ) == 0: # one of the a or b is a root for the function return a elif function(UpperCamelCase ) == 0: return b elif ( function(UpperCamelCase ) * function(UpperCamelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError("""could not find root in given interval.""" ) else: lowerCAmelCase__ : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(UpperCamelCase ) == 0: return mid elif function(UpperCamelCase ) * function(UpperCamelCase ) < 0: lowerCAmelCase__ : Optional[Any] = mid else: lowerCAmelCase__ : Union[str, Any] = mid lowerCAmelCase__ : Any = start + (end - start) / 2.0 return mid def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1000)) import doctest doctest.testmod()
37
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
0
from collections.abc import Sequence from queue import Queue class _SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Any , __lowerCamelCase : str=None , __lowerCamelCase : Optional[Any]=None ): UpperCamelCase :List[Any] = start UpperCamelCase :str = end UpperCamelCase :Tuple = val UpperCamelCase :Tuple = (start + end) // 2 UpperCamelCase :str = left UpperCamelCase :Optional[int] = right def __repr__( self : List[str] ): return F"""SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})""" class _SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , __lowerCamelCase : Sequence , __lowerCamelCase : List[Any] ): UpperCamelCase :int = collection UpperCamelCase :List[Any] = function if self.collection: UpperCamelCase :Tuple = self._build_tree(0 , len(__lowerCamelCase ) - 1 ) def _A ( self : Any , __lowerCamelCase : List[Any] , __lowerCamelCase : int ): self._update_tree(self.root , __lowerCamelCase , __lowerCamelCase ) def _A ( self : Any , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple ): return self._query_range(self.root , __lowerCamelCase , __lowerCamelCase ) def _A ( self : Tuple , __lowerCamelCase : str , __lowerCamelCase : List[str] ): if start == end: return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.collection[start] ) UpperCamelCase :Dict = (start + end) // 2 UpperCamelCase :List[str] = self._build_tree(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :Union[str, Any] = self._build_tree(mid + 1 , __lowerCamelCase ) return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.fn(left.val , right.val ) , __lowerCamelCase , __lowerCamelCase ) def _A ( self : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str] ): if node.start == i and node.end == i: UpperCamelCase :List[Any] = val return if i <= node.mid: self._update_tree(node.left , __lowerCamelCase , __lowerCamelCase ) else: self._update_tree(node.right , __lowerCamelCase , __lowerCamelCase ) UpperCamelCase :Union[str, Any] = self.fn(node.left.val , node.right.val ) def _A ( self : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : int ): if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , __lowerCamelCase , __lowerCamelCase ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , __lowerCamelCase , node.mid ) , self._query_range(node.right , node.mid + 1 , __lowerCamelCase ) , ) else: # range in right child tree return self._query_range(node.right , __lowerCamelCase , __lowerCamelCase ) def _A ( self : Optional[Any] ): if self.root is not None: UpperCamelCase :Union[str, Any] = Queue() queue.put(self.root ) while not queue.empty(): UpperCamelCase :Optional[int] = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 50) UpperCAmelCase_ : Optional[int] = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
38
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def lowercase__ ( __lowercase : Tuple ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = emb.weight.shape __UpperCamelCase = nn.Linear(__lowercase , __lowercase , bias=__lowercase ) __UpperCamelCase = emb.weight.data return lin_layer def lowercase__ ( __lowercase : int , __lowercase : List[str]="facebook/mbart-large-en-ro" , __lowercase : str=False , __lowercase : List[Any]=False ) -> int: """simple docstring""" __UpperCamelCase = torch.load(__lowercase , map_location='cpu' )['model'] remove_ignore_keys_(__lowercase ) __UpperCamelCase = state_dict['encoder.embed_tokens.weight'].shape[0] __UpperCamelCase = MBartConfig.from_pretrained(__lowercase , vocab_size=__lowercase ) if mbart_aa and finetuned: __UpperCamelCase = 'relu' __UpperCamelCase = state_dict['decoder.embed_tokens.weight'] __UpperCamelCase = MBartForConditionalGeneration(__lowercase ) model.model.load_state_dict(__lowercase ) if finetuned: __UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": a__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') a__ : Union[str, Any] =parser.parse_args() a__ : str =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
53
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __A ( __lowerCAmelCase , __lowerCAmelCase=False )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('head' ): _UpperCAmelCase = 'segformer.encoder.' + key if key.startswith('backbone' ): _UpperCAmelCase = key.replace('backbone' , 'segformer.encoder' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 _UpperCAmelCase = key[key.find('patch_embed' ) + len('patch_embed' )] _UpperCAmelCase = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(__lowerCAmelCase )-1}""" ) if "norm" in key: _UpperCAmelCase = key.replace('norm' , 'layer_norm' ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 _UpperCAmelCase = key[key.find('segformer.encoder.layer_norm' ) + len('segformer.encoder.layer_norm' )] _UpperCAmelCase = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(__lowerCAmelCase )-1}""" ) if "layer_norm1" in key: _UpperCAmelCase = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: _UpperCAmelCase = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 _UpperCAmelCase = key[key.find('block' ) + len('block' )] _UpperCAmelCase = key.replace(F"""block{idx}""" , F"""block.{int(__lowerCAmelCase )-1}""" ) if "attn.q" in key: _UpperCAmelCase = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: _UpperCAmelCase = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: _UpperCAmelCase = key.replace('attn' , 'attention.self' ) if "fc1" in key: _UpperCAmelCase = key.replace('fc1' , 'dense1' ) if "fc2" in key: _UpperCAmelCase = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: _UpperCAmelCase = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: _UpperCAmelCase = key.replace('linear_fuse.conv' , 'linear_fuse' ) _UpperCAmelCase = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 _UpperCAmelCase = key[key.find('linear_c' ) + len('linear_c' )] _UpperCAmelCase = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(__lowerCAmelCase )-1}""" ) if key.startswith('head' ): _UpperCAmelCase = key.replace('head' , 'classifier' ) _UpperCAmelCase = value return new_state_dict def __A ( __lowerCAmelCase , __lowerCAmelCase )-> List[str]: """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) _UpperCAmelCase = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) _UpperCAmelCase = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict _UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] _UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] _UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] _UpperCAmelCase = kv_bias[ config.hidden_sizes[i] : ] def __A ( )-> Union[str, Any]: """simple docstring""" _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return image @torch.no_grad() def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> Tuple: """simple docstring""" _UpperCAmelCase = SegformerConfig() _UpperCAmelCase = False # set attributes based on model_name _UpperCAmelCase = 'huggingface/label-files' if "segformer" in model_name: _UpperCAmelCase = model_name[len('segformer.' ) : len('segformer.' ) + 2] if "ade" in model_name: _UpperCAmelCase = 150 _UpperCAmelCase = 'ade20k-id2label.json' _UpperCAmelCase = (1, 150, 128, 128) elif "city" in model_name: _UpperCAmelCase = 19 _UpperCAmelCase = 'cityscapes-id2label.json' _UpperCAmelCase = (1, 19, 128, 128) else: raise ValueError(F"""Model {model_name} not supported""" ) elif "mit" in model_name: _UpperCAmelCase = True _UpperCAmelCase = model_name[4:6] _UpperCAmelCase = 1_000 _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = (1, 1_000) else: raise ValueError(F"""Model {model_name} not supported""" ) # set config attributes _UpperCAmelCase = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": _UpperCAmelCase = [64, 128, 320, 512] _UpperCAmelCase = 256 elif size == "b2": _UpperCAmelCase = [64, 128, 320, 512] _UpperCAmelCase = 768 _UpperCAmelCase = [3, 4, 6, 3] elif size == "b3": _UpperCAmelCase = [64, 128, 320, 512] _UpperCAmelCase = 768 _UpperCAmelCase = [3, 4, 18, 3] elif size == "b4": _UpperCAmelCase = [64, 128, 320, 512] _UpperCAmelCase = 768 _UpperCAmelCase = [3, 8, 27, 3] elif size == "b5": _UpperCAmelCase = [64, 128, 320, 512] _UpperCAmelCase = 768 _UpperCAmelCase = [3, 6, 40, 3] else: raise ValueError(F"""Size {size} not supported""" ) # load image processor (only resize + normalize) _UpperCAmelCase = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=__lowerCAmelCase , align=__lowerCAmelCase , do_random_crop=__lowerCAmelCase ) # prepare image _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=__lowerCAmelCase , return_tensors='pt' ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict if encoder_only: _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location=torch.device('cpu' ) ) else: _UpperCAmelCase = torch.load(__lowerCAmelCase , map_location=torch.device('cpu' ) )['state_dict'] # rename keys _UpperCAmelCase = rename_keys(__lowerCAmelCase , encoder_only=__lowerCAmelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(__lowerCAmelCase , __lowerCAmelCase ) # create HuggingFace model and load state dict if encoder_only: _UpperCAmelCase = False _UpperCAmelCase = SegformerForImageClassification(__lowerCAmelCase ) else: _UpperCAmelCase = SegformerForSemanticSegmentation(__lowerCAmelCase ) model.load_state_dict(__lowerCAmelCase ) model.eval() # forward pass _UpperCAmelCase = model(__lowerCAmelCase ) _UpperCAmelCase = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-7.58_20, -8.72_31, -8.32_15], [-8.06_00, -10.35_29, -10.03_04], [-7.52_08, -9.41_03, -9.62_39]], [[-12.69_18, -13.89_94, -13.71_37], [-13.31_96, -15.75_23, -15.47_89], [-12.93_43, -14.87_57, -14.96_89]], [[-11.19_11, -11.94_21, -11.32_43], [-11.33_42, -13.68_39, -13.35_81], [-10.39_09, -12.18_32, -12.48_58]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-11.81_73, -14.38_50, -16.31_28], [-14.56_48, -16.58_04, -18.65_68], [-14.72_23, -15.73_87, -18.42_18]], [[-15.72_90, -17.91_71, -19.44_23], [-18.31_05, -19.94_48, -21.46_61], [-17.92_96, -18.64_97, -20.79_10]], [[-15.07_83, -17.03_36, -18.27_89], [-16.87_71, -18.68_70, -20.16_12], [-16.24_54, -17.14_26, -19.50_55]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-9.08_78, -10.20_81, -10.18_91], [-9.31_44, -10.79_41, -10.98_43], [-9.22_94, -10.38_55, -10.57_04]], [[-12.23_16, -13.90_68, -13.61_02], [-12.91_61, -14.37_02, -14.32_35], [-12.52_33, -13.71_74, -13.79_32]], [[-14.62_75, -15.24_90, -14.97_27], [-14.34_00, -15.96_87, -16.28_27], [-14.14_84, -15.40_33, -15.89_37]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-12.31_44, -13.24_47, -14.08_02], [-13.36_14, -14.58_16, -15.61_17], [-13.33_40, -14.44_33, -16.22_19]], [[-19.27_81, -20.41_28, -20.75_06], [-20.61_53, -21.65_66, -22.09_98], [-19.98_00, -21.04_30, -22.14_94]], [[-18.87_39, -19.78_04, -21.18_34], [-20.12_33, -21.67_65, -23.29_44], [-20.03_15, -21.26_41, -23.69_44]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": _UpperCAmelCase = torch.tensor( [ [[-9.55_24, -12.08_35, -11.73_48], [-10.52_29, -13.64_46, -14.56_62], [-9.58_42, -12.88_51, -13.94_14]], [[-15.34_32, -17.53_23, -17.08_18], [-16.33_30, -18.92_55, -19.21_01], [-15.13_40, -17.78_48, -18.39_71]], [[-12.60_72, -14.94_86, -14.66_31], [-13.76_29, -17.09_07, -17.77_45], [-12.78_99, -16.16_95, -17.16_71]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-11.92_95, -13.40_57, -14.81_06], [-13.34_31, -14.81_79, -15.37_81], [-14.28_36, -15.59_42, -16.15_88]], [[-11.49_06, -12.80_67, -13.65_64], [-13.11_89, -14.05_00, -14.15_43], [-13.87_48, -14.51_36, -14.87_89]], [[0.53_74, 0.10_67, -0.47_42], [0.11_41, -0.22_55, -0.70_99], [-0.30_00, -0.59_24, -1.31_05]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-7.82_17, -9.87_67, -10.17_17], [-9.44_38, -10.90_58, -11.40_47], [-9.79_39, -12.34_95, -12.10_79]], [[-7.15_14, -9.53_36, -10.08_60], [-9.77_76, -11.68_22, -11.84_39], [-10.14_11, -12.76_55, -12.89_72]], [[0.30_21, 0.08_05, -0.23_10], [-0.03_28, -0.16_05, -0.27_14], [-0.14_08, -0.54_77, -0.69_76]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": _UpperCAmelCase = torch.tensor( [ [ [-1.1372E01, -1.2787E01, -1.3477E01], [-1.2536E01, -1.4194E01, -1.4409E01], [-1.3217E01, -1.4888E01, -1.5327E01], ], [ [-1.4791E01, -1.7122E01, -1.8277E01], [-1.7163E01, -1.9192E01, -1.9533E01], [-1.7897E01, -1.9991E01, -2.0315E01], ], [ [7.6723E-01, 4.1921E-01, -7.7878E-02], [4.7772E-01, 9.5557E-03, -2.8082E-01], [3.6032E-01, -2.4826E-01, -5.1168E-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": _UpperCAmelCase = torch.tensor( [ [[-9.49_59, -11.30_87, -11.74_79], [-11.00_25, -12.65_40, -12.33_19], [-11.40_64, -13.04_87, -12.99_05]], [[-9.89_05, -11.30_84, -12.08_54], [-11.17_26, -12.76_98, -12.95_83], [-11.59_85, -13.32_78, -14.17_74]], [[0.22_13, 0.01_92, -0.24_66], [-0.17_31, -0.42_13, -0.48_74], [-0.31_26, -0.65_41, -1.13_89]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-16.09_76, -16.48_56, -17.39_62], [-16.62_34, -19.03_42, -19.76_85], [-16.09_00, -18.06_61, -19.11_80]], [[-18.47_50, -18.84_88, -19.50_74], [-19.40_30, -22.15_70, -22.59_77], [-19.11_91, -20.84_86, -22.37_83]], [[-4.51_78, -5.50_37, -6.51_09], [-5.08_84, -7.21_74, -8.03_34], [-4.41_56, -5.81_17, -7.29_70]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-14.20_81, -14.47_32, -14.19_77], [-14.58_67, -16.44_23, -16.63_56], [-13.44_41, -14.96_85, -16.86_96]], [[-14.45_76, -14.70_73, -15.04_51], [-15.08_16, -17.62_37, -17.98_73], [-14.42_13, -16.01_99, -18.59_92]], [[-4.73_49, -4.95_88, -5.09_66], [-4.32_10, -6.93_25, -7.25_91], [-3.43_12, -4.74_84, -7.19_17]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-11.77_37, -11.95_26, -11.32_73], [-13.66_92, -14.45_74, -13.88_78], [-13.89_37, -14.69_24, -15.93_45]], [[-14.67_06, -14.53_30, -14.13_06], [-16.15_02, -16.81_80, -16.42_69], [-16.83_38, -17.89_39, -20.17_46]], [[1.04_91, 0.82_89, 1.03_10], [1.10_44, 0.52_19, 0.80_55], [1.08_99, 0.69_26, 0.55_90]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": _UpperCAmelCase = torch.tensor( [ [[-12.56_41, -13.47_77, -13.06_84], [-13.95_87, -15.89_83, -16.65_57], [-13.31_09, -15.73_50, -16.31_41]], [[-14.70_74, -15.43_52, -14.59_44], [-16.63_53, -18.16_63, -18.61_20], [-15.17_02, -18.03_29, -18.15_47]], [[-1.79_90, -2.09_51, -1.77_84], [-2.63_97, -3.82_45, -3.96_86], [-1.52_64, -2.81_26, -2.93_16]], ] ) else: _UpperCAmelCase = logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , __lowerCAmelCase , atol=1E-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) model.save_pretrained(__lowerCAmelCase ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''segformer.b0.512x512.ade.160k''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) _a = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
39
'''simple docstring''' import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Any , __A : Dict , __A : str , __A : List[Any]=1_0_2_4 , __A : Tuple=1_0_2_4 , __A : str=3.6 ): __UpperCamelCase = tokenizer __UpperCamelCase = tokenizer.bos_token_id __UpperCamelCase = dataset __UpperCamelCase = seq_length __UpperCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self : Any ): __UpperCamelCase = iter(self.dataset ) __UpperCamelCase = True while more_examples: __UpperCamelCase , __UpperCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__A )['content'] ) buffer_len += len(buffer[-1] ) except StopIteration: __UpperCamelCase = False break __UpperCamelCase = tokenizer(__A , truncation=__A )['input_ids'] __UpperCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__A ) , self.seq_length ): __UpperCamelCase = all_token_ids[i : i + self.seq_length] if len(__A ) == self.seq_length: yield torch.tensor(__A ) def lowercase__ ( __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = {'streaming': True} __UpperCamelCase = load_dataset(args.dataset_name , split='train' , **__lowercase ) __UpperCamelCase = ConstantLengthDataset(__lowercase , __lowercase , seq_length=args.seq_length ) __UpperCamelCase = DataLoader(__lowercase , batch_size=args.batch_size ) return eval_dataloader def lowercase__ ( __lowercase : Tuple ) -> Optional[Any]: """simple docstring""" model.eval() __UpperCamelCase = [] for step, batch in enumerate(__lowercase ): with torch.no_grad(): __UpperCamelCase = model(__lowercase , labels=__lowercase ) __UpperCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(__lowercase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __UpperCamelCase = torch.mean(torch.cat(__lowercase ) ) try: __UpperCamelCase = torch.exp(__lowercase ) except OverflowError: __UpperCamelCase = float('inf' ) return loss.item(), perplexity.item() # Setup Accelerator a__ : int =Accelerator() # Parse configuration a__ : Dict =HfArgumentParser(EvaluationArguments) a__ : Union[str, Any] =parser.parse_args() set_seed(args.seed) # Logging a__ : List[Any] =logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer a__ : Union[str, Any] =AutoModelForCausalLM.from_pretrained(args.model_ckpt) a__ : List[Any] =AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a__ : Union[str, Any] =create_dataloader(args) # Prepare everything with our `accelerator`. a__ , a__ : List[str] =accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') a__ , a__ : Any =evaluate(args) logger.info(f'loss/eval: {eval_loss}, perplexity: {perplexity}')
53
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowercase = logging.get_logger(__name__) class _A ( _a ): """simple docstring""" def __init__( self : List[Any] , *__UpperCAmelCase : Optional[Any] , **__UpperCAmelCase : Optional[Any]): warnings.warn( "The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use BeitImageProcessor instead." , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase)
40
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging a__ : Any =logging.get_logger(__name__) a__ : Optional[Any] ={ '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict ="gpt_neo" SCREAMING_SNAKE_CASE_ : Optional[int] =["past_key_values"] SCREAMING_SNAKE_CASE_ : List[Any] ={"num_attention_heads": "num_heads", "num_hidden_layers": "num_layers"} def __init__( self : Union[str, Any] , __A : Union[str, Any]=5_0_2_5_7 , __A : Any=2_0_4_8 , __A : Optional[Any]=2_0_4_8 , __A : Any=2_4 , __A : Union[str, Any]=[[["global", "local"], 1_2]] , __A : str=1_6 , __A : Optional[int]=None , __A : Union[str, Any]=2_5_6 , __A : Any="gelu_new" , __A : Dict=0.0 , __A : Optional[int]=0.0 , __A : int=0.0 , __A : List[str]=0.1 , __A : Any=1e-5 , __A : int=0.02 , __A : List[str]=True , __A : Tuple=5_0_2_5_6 , __A : Optional[Any]=5_0_2_5_6 , **__A : Optional[Any] , ): __UpperCamelCase = vocab_size __UpperCamelCase = max_position_embeddings __UpperCamelCase = hidden_size __UpperCamelCase = num_layers __UpperCamelCase = num_heads __UpperCamelCase = intermediate_size __UpperCamelCase = window_size __UpperCamelCase = activation_function __UpperCamelCase = resid_dropout __UpperCamelCase = embed_dropout __UpperCamelCase = attention_dropout __UpperCamelCase = classifier_dropout __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = initializer_range __UpperCamelCase = use_cache __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id __UpperCamelCase = attention_types __UpperCamelCase = self.expand_attention_types_params(__A ) if len(self.attention_layers ) != self.num_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.attention_layers)` == `config.num_layers` ' f'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' f'''`config.num_layers = {self.num_layers}`. ''' '`config.attention_layers` is prepared using `config.attention_types`. ' 'Please verify the value of `config.attention_types` argument.' ) super().__init__(bos_token_id=__A , eos_token_id=__A , **__A ) @staticmethod def _lowerCamelCase ( __A : Tuple ): __UpperCamelCase = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase__ ( __lowercase : Tuple , __lowercase : Any , __lowercase : Union[str, Any] , __lowercase : List[str] ) -> Any: """simple docstring""" import torch __UpperCamelCase = input.size() __UpperCamelCase = len(__lowercase ) __UpperCamelCase = shape[dimension] __UpperCamelCase = torch.arange(0 , __lowercase , __lowercase ) __UpperCamelCase = torch.div(sizedim - size , __lowercase , rounding_mode='floor' ) + 1 __UpperCamelCase = torch.arange(__lowercase ) + low_indices[:min_length][:, None] __UpperCamelCase = [slice(__lowercase )] * rank __UpperCamelCase = indices __UpperCamelCase = input[s] __UpperCamelCase = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" import torch __UpperCamelCase = torch.arange(1 , __lowercase ) __UpperCamelCase = torch.remainder(__lowercase , __lowercase ) __UpperCamelCase = remainders == 0 __UpperCamelCase = candidates[divisor_indices] __UpperCamelCase = torch.max(__lowercase ) return largest_divisor, torch.div(__lowercase , __lowercase , rounding_mode='floor' ) class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(__A , direction='inputs' ) __UpperCamelCase = {0: 'batch', 1: 'past_sequence + sequence'} else: __UpperCamelCase = {0: 'batch', 1: 'sequence'} return common_inputs @property def _lowerCamelCase ( self : int ): return self._config.num_heads def _lowerCamelCase ( self : List[str] , __A : PreTrainedTokenizer , __A : int = -1 , __A : int = -1 , __A : bool = False , __A : Optional[TensorType] = None , ): __UpperCamelCase = super(__A , self ).generate_dummy_inputs( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) # We need to order the input in the way they appears in the forward() __UpperCamelCase = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __UpperCamelCase , __UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values __UpperCamelCase = seqlen + 2 __UpperCamelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __UpperCamelCase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(self.num_layers ) ] __UpperCamelCase = common_inputs['attention_mask'] if self.use_past: __UpperCamelCase = ordered_inputs['attention_mask'].dtype __UpperCamelCase = torch.cat( [ordered_inputs['attention_mask'], torch.ones(__A , __A , dtype=__A )] , dim=1 ) return ordered_inputs @property def _lowerCamelCase ( self : Dict ): return 1_3
53
0
'''simple docstring''' _A : Any ={'''a''': ['''c''', '''b'''], '''b''': ['''d''', '''e'''], '''c''': [], '''d''': [], '''e''': []} _A : Optional[Any] =['''a''', '''b''', '''c''', '''d''', '''e'''] def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: lowerCamelCase__ : str = start # add current to visited visited.append(UpperCamelCase ) lowerCamelCase__ : Optional[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: lowerCamelCase__ : str = topological_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # if all neighbors visited add current to sort sort.append(UpperCamelCase ) # if all vertices haven't been visited select a new one to visit if len(UpperCamelCase ) != len(UpperCamelCase ): for vertice in vertices: if vertice not in visited: lowerCamelCase__ : Union[str, Any] = topological_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # return sort return sort if __name__ == "__main__": _A : Optional[Any] =topological_sort('''a''', [], []) print(sort)
41
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="naver-clova-ix/donut-base-finetuned-docvqa" SCREAMING_SNAKE_CASE_ : Dict =( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) SCREAMING_SNAKE_CASE_ : List[str] ="document_qa" SCREAMING_SNAKE_CASE_ : Union[str, Any] =AutoProcessor SCREAMING_SNAKE_CASE_ : Union[str, Any] =VisionEncoderDecoderModel SCREAMING_SNAKE_CASE_ : List[Any] =["image", "text"] SCREAMING_SNAKE_CASE_ : Any =["text"] def __init__( self : Optional[int] , *__A : List[str] , **__A : List[Any] ): if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.' ) super().__init__(*__A , **__A ) def _lowerCamelCase ( self : Any , __A : "Image" , __A : str ): __UpperCamelCase = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' __UpperCamelCase = task_prompt.replace('{user_input}' , __A ) __UpperCamelCase = self.pre_processor.tokenizer( __A , add_special_tokens=__A , return_tensors='pt' ).input_ids __UpperCamelCase = self.pre_processor(__A , return_tensors='pt' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _lowerCamelCase ( self : Union[str, Any] , __A : Optional[Any] ): return self.model.generate( inputs['pixel_values'].to(self.device ) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=__A , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=__A , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=__A , ).sequences def _lowerCamelCase ( self : Tuple , __A : List[Any] ): __UpperCamelCase = self.pre_processor.batch_decode(__A )[0] __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '' ) __UpperCamelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '' ) __UpperCamelCase = re.sub(R'<.*?>' , '' , __A , count=1 ).strip() # remove first task start token __UpperCamelCase = self.pre_processor.tokenajson(__A ) return sequence["answer"]
53
0
'''simple docstring''' import os lowercase : Optional[int] = {"I": 1, "V": 5, "X": 10, "L": 50, "C": 100, "D": 500, "M": 1000} def SCREAMING_SNAKE_CASE__ ( __A ) -> int: _snake_case = 0 _snake_case = 0 while index < len(__A ) - 1: _snake_case = SYMBOLS[numerals[index]] _snake_case = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = '' _snake_case = num // 1_000 numerals += m_count * "M" num %= 1_000 _snake_case = num // 100 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 100 _snake_case = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def SCREAMING_SNAKE_CASE__ ( __A = "/p089_roman.txt" ) -> int: _snake_case = 0 with open(os.path.dirname(__A ) + roman_numerals_filename ) as filea: _snake_case = filea.readlines() for line in lines: _snake_case = line.strip() _snake_case = parse_roman_numerals(__A ) _snake_case = generate_roman_numerals(__A ) savings += len(__A ) - len(__A ) return savings if __name__ == "__main__": print(F'''{solution() = }''')
42
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
53
0
from __future__ import annotations def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return [ord(SCREAMING_SNAKE_CASE ) - 96 for elem in plain] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return "".join(chr(elem + 96 ) for elem in encoded ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :List[Any] = encode(input('''-> ''' ).strip().lower() ) print('''Encoded: ''' , SCREAMING_SNAKE_CASE ) print('''Decoded:''' , decode(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main()
43
'''simple docstring''' import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def lowercase__ ( __lowercase : Features ) -> Optional[int]: """simple docstring""" __UpperCamelCase = np.inf def set_batch_size(__lowercase : FeatureType ) -> None: nonlocal batch_size if isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__lowercase , __lowercase ): __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__lowercase , __lowercase ) and feature.dtype == "binary": __UpperCamelCase = min(__lowercase , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__lowercase , __lowercase ) return None if batch_size is np.inf else batch_size class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : List[str] , __A : NestedDataStructureLike[PathLike] , __A : Optional[NamedSplit] = None , __A : Optional[Features] = None , __A : str = None , __A : bool = False , __A : bool = False , __A : Optional[int] = None , **__A : Dict , ): super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) __UpperCamelCase = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} __UpperCamelCase = _PACKAGED_DATASETS_MODULES['parquet'][1] __UpperCamelCase = Parquet( cache_dir=__A , data_files=__A , features=__A , hash=__A , **__A , ) def _lowerCamelCase ( self : Optional[int] ): # Build iterable dataset if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset class snake_case : """simple docstring""" def __init__( self : List[str] , __A : Dataset , __A : Union[PathLike, BinaryIO] , __A : Optional[int] = None , **__A : Dict , ): __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size or get_writer_batch_size(dataset.features ) __UpperCamelCase = parquet_writer_kwargs def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , 'wb+' ) as buffer: __UpperCamelCase = self._write(file_obj=__A , batch_size=__A , **self.parquet_writer_kwargs ) else: __UpperCamelCase = self._write(file_obj=self.path_or_buf , batch_size=__A , **self.parquet_writer_kwargs ) return written def _lowerCamelCase ( self : List[str] , __A : BinaryIO , __A : int , **__A : List[str] ): __UpperCamelCase = 0 __UpperCamelCase = parquet_writer_kwargs.pop('path_or_buf' , __A ) __UpperCamelCase = self.dataset.features.arrow_schema __UpperCamelCase = pq.ParquetWriter(__A , schema=__A , **__A ) for offset in logging.tqdm( range(0 , len(self.dataset ) , __A ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating parquet from Arrow format' , ): __UpperCamelCase = query_table( table=self.dataset._data , key=slice(__A , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(__A ) written += batch.nbytes writer.close() return written
53
0
"""simple docstring""" _a : Dict = 0 # The first color of the flag. _a : Union[str, Any] = 1 # The second color of the flag. _a : Dict = 2 # The third color of the flag. _a : Union[str, Any] = (red, white, blue) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : list ) -> list: if not sequence: return [] if len(_lowerCamelCase ) == 1: return list(_lowerCamelCase ) _lowerCAmelCase : List[Any] = 0 _lowerCAmelCase : Union[str, Any] = len(_lowerCamelCase ) - 1 _lowerCAmelCase : Tuple = 0 while mid <= high: if sequence[mid] == colors[0]: _lowerCAmelCase , _lowerCAmelCase : Any = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: _lowerCAmelCase , _lowerCAmelCase : Any = sequence[high], sequence[mid] high -= 1 else: _lowerCAmelCase : Dict = f"The elements inside the sequence must contains only {colors} values" raise ValueError(_lowerCamelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod() _a : int = input('Enter numbers separated by commas:\n').strip() _a : Any = [int(item.strip()) for item in user_input.split(',')] print(F"""{dutch_national_flag_sort(unsorted)}""")
44
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def lowercase__ ( __lowercase : SplitDict ) -> int: """simple docstring""" __UpperCamelCase = split_dict._to_yaml_list() assert len(__lowercase ) == len(__lowercase ) __UpperCamelCase = SplitDict._from_yaml_list(__lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __UpperCamelCase = None # the split name of split_dict takes over the name of the split info object __UpperCamelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=__lowercase ), SplitInfo(dataset_name='my_dataset' )] ) def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" __UpperCamelCase = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
53
0
"""simple docstring""" import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __UpperCAmelCase ( self ): __a = tempfile.mkdtemp() __a = 8 # DPR tok __a = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __a = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(_a , exist_ok=_a ) __a = os.path.join(_a , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __a = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __a = dict(zip(_a , range(len(_a ) ) ) ) __a = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __a = {'''unk_token''': '''<unk>'''} __a = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(_a , exist_ok=_a ) __a = os.path.join(_a , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __a = os.path.join(_a , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_a ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(_a ) ) def __UpperCAmelCase ( self ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def __UpperCAmelCase ( self ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def __UpperCAmelCase ( self ): shutil.rmtree(self.tmpdirname ) @require_tokenizers def __UpperCAmelCase ( self ): __a = os.path.join(self.tmpdirname , '''rag_tokenizer''' ) __a = RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) __a = RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(_a ) rag_tokenizer.save_pretrained(_a ) __a = RagTokenizer.from_pretrained(_a , config=_a ) self.assertIsInstance(new_rag_tokenizer.question_encoder , _a ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , _a ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __UpperCAmelCase ( self ): __a = RagTokenizer.from_pretrained('''facebook/rag-token-nq''' ) __a = [ '''who got the first nobel prize in physics''', '''when is the next deadpool movie being released''', '''which mode is used for short wave broadcast service''', '''who is the owner of reading football club''', '''when is the next scandal episode coming out''', '''when is the last time the philadelphia won the superbowl''', '''what is the most current adobe flash player version''', '''how many episodes are there in dragon ball z''', '''what is the first step in the evolution of the eye''', '''where is gall bladder situated in human body''', '''what is the main mineral in lithium batteries''', '''who is the president of usa right now''', '''where do the greasers live in the outsiders''', '''panda is a national animal of which country''', '''what is the name of manchester united stadium''', ] __a = tokenizer(_a ) self.assertIsNotNone(_a ) @slow def __UpperCAmelCase ( self ): __a = RagTokenizer.from_pretrained('''facebook/rag-sequence-nq''' ) __a = [ '''who got the first nobel prize in physics''', '''when is the next deadpool movie being released''', '''which mode is used for short wave broadcast service''', '''who is the owner of reading football club''', '''when is the next scandal episode coming out''', '''when is the last time the philadelphia won the superbowl''', '''what is the most current adobe flash player version''', '''how many episodes are there in dragon ball z''', '''what is the first step in the evolution of the eye''', '''where is gall bladder situated in human body''', '''what is the main mineral in lithium batteries''', '''who is the president of usa right now''', '''where do the greasers live in the outsiders''', '''panda is a national animal of which country''', '''what is the name of manchester united stadium''', ] __a = tokenizer(_a ) self.assertIsNotNone(_a )
45
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[str] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Any =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys a__ : str =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
"""simple docstring""" def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
46
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a__ : str =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str =["input_features", "attention_mask"] def __init__( self : Union[str, Any] , __A : Optional[int]=8_0 , __A : Tuple=1_6_0_0_0 , __A : Optional[Any]=8_0 , __A : Any=0.0 , __A : Any=True , __A : List[str]=True , __A : str=True , **__A : List[Any] , ): super().__init__(feature_size=__A , sampling_rate=__A , padding_value=__A , **__A ) __UpperCamelCase = num_mel_bins __UpperCamelCase = do_ceptral_normalize __UpperCamelCase = normalize_means __UpperCamelCase = normalize_vars __UpperCamelCase = True def _lowerCamelCase ( self : Union[str, Any] , __A : np.ndarray , ): __UpperCamelCase = waveform * (2**1_5) # Kaldi compliance: 16-bit signed integers __UpperCamelCase = torch.from_numpy(__A ).unsqueeze(0 ) __UpperCamelCase = ta_kaldi.fbank(__A , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( __A : np.ndarray , __A : int , __A : Optional[bool] = True , __A : Optional[bool] = True , __A : float = 0.0 , ): # make sure we normalize float32 arrays if normalize_means: __UpperCamelCase = x[:input_length].mean(axis=0 ) __UpperCamelCase = np.subtract(__A , __A ) if normalize_vars: __UpperCamelCase = x[:input_length].std(axis=0 ) __UpperCamelCase = np.divide(__A , __A ) if input_length < x.shape[0]: __UpperCamelCase = padding_value # make sure array is in float32 __UpperCamelCase = x.astype(np.floataa ) return x def _lowerCamelCase ( self : int , __A : List[np.ndarray] , __A : Optional[np.ndarray] = None ): __UpperCamelCase = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(__A , __A , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(__A , __A ) ] def __call__( self : List[Any] , __A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __A : Union[bool, str, PaddingStrategy] = False , __A : Optional[int] = None , __A : bool = False , __A : Optional[int] = None , __A : Optional[Union[str, TensorType]] = None , __A : Optional[int] = None , __A : Optional[bool] = None , **__A : Dict , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) __UpperCamelCase = isinstance(__A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __UpperCamelCase = is_batched_numpy or ( isinstance(__A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(__A , np.ndarray ): __UpperCamelCase = np.asarray(__A , dtype=np.floataa ) elif isinstance(__A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCamelCase = [raw_speech] # extract fbank features __UpperCamelCase = [self._extract_fbank_features(__A ) for waveform in raw_speech] # convert into correct format for padding __UpperCamelCase = BatchFeature({'input_features': features} ) __UpperCamelCase = self.pad( __A , padding=__A , max_length=__A , truncation=__A , pad_to_multiple_of=__A , return_attention_mask=__A , **__A , ) # make sure list is in array format __UpperCamelCase = padded_inputs.get('input_features' ) if isinstance(input_features[0] , __A ): __UpperCamelCase = [np.asarray(__A , dtype=np.floataa ) for feature in input_features] __UpperCamelCase = padded_inputs.get('attention_mask' ) if attention_mask is not None: __UpperCamelCase = [np.asarray(__A , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __UpperCamelCase = ( np.array(__A , dtype=np.intaa ) if self._get_padding_strategies(__A , max_length=__A ) is not PaddingStrategy.DO_NOT_PAD else None ) __UpperCamelCase = self.normalize( padded_inputs['input_features'] , attention_mask=__A ) if return_tensors is not None: __UpperCamelCase = padded_inputs.convert_to_tensors(__A ) return padded_inputs
53
0
'''simple docstring''' import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem lowerCamelCase : Union[str, Any] = importlib.util.find_spec("s3fs") is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 lowerCamelCase : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(f'''A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.''') fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def _lowerCAmelCase ( _UpperCamelCase : str ) -> str: """simple docstring""" if "://" in dataset_path: _SCREAMING_SNAKE_CASE =dataset_path.split('://' )[1] return dataset_path def _lowerCAmelCase ( _UpperCamelCase : fsspec.AbstractFileSystem ) -> bool: """simple docstring""" if fs is not None and fs.protocol != "file": return True else: return False def _lowerCAmelCase ( _UpperCamelCase : fsspec.AbstractFileSystem , _UpperCamelCase : str , _UpperCamelCase : str ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =not is_remote_filesystem(_UpperCamelCase ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(_UpperCamelCase ) , fs._strip_protocol(_UpperCamelCase ) ) else: fs.mv(_UpperCamelCase , _UpperCamelCase , recursive=_UpperCamelCase ) def _lowerCAmelCase ( ) -> None: """simple docstring""" if hasattr(fsspec.asyn , 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: _SCREAMING_SNAKE_CASE =None _SCREAMING_SNAKE_CASE =None _SCREAMING_SNAKE_CASE =threading.Lock()
47
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : List[Any] =logging.get_logger(__name__) a__ : List[Any] ={ '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_text_model" def __init__( self : str , __A : List[Any]=2_5_0_0_0_2 , __A : Any=1_0_2_4 , __A : int=2_4 , __A : Dict=1_6 , __A : Optional[Any]=4_0_9_6 , __A : Union[str, Any]="gelu" , __A : Dict=0.1 , __A : Dict=0.1 , __A : List[str]=5_1_4 , __A : Optional[int]=1 , __A : int=0.02 , __A : Optional[Any]=0.02 , __A : Optional[Any]=1e-05 , __A : Dict=1 , __A : List[Any]=0 , __A : int=2 , __A : Tuple="absolute" , __A : Optional[Any]=True , __A : Optional[int]=7_6_8 , **__A : List[str] , ): super().__init__(pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = use_cache __UpperCamelCase = project_dim class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="altclip_vision_model" def __init__( self : List[Any] , __A : Union[str, Any]=7_6_8 , __A : Optional[int]=3_0_7_2 , __A : Optional[Any]=5_1_2 , __A : Tuple=1_2 , __A : Union[str, Any]=1_2 , __A : Optional[int]=3 , __A : Dict=2_2_4 , __A : Tuple=3_2 , __A : str="quick_gelu" , __A : Dict=1e-5 , __A : Optional[int]=0.0 , __A : List[Any]=0.02 , __A : int=1.0 , **__A : Optional[int] , ): super().__init__(**__A ) __UpperCamelCase = hidden_size __UpperCamelCase = intermediate_size __UpperCamelCase = projection_dim __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = num_channels __UpperCamelCase = patch_size __UpperCamelCase = image_size __UpperCamelCase = initializer_range __UpperCamelCase = initializer_factor __UpperCamelCase = attention_dropout __UpperCamelCase = layer_norm_eps __UpperCamelCase = hidden_act @classmethod def _lowerCamelCase ( cls : Optional[Any] , __A : Union[str, os.PathLike] , **__A : Optional[Any] ): cls._set_token_in_kwargs(__A ) __UpperCamelCase , __UpperCamelCase = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('model_type' ) == "altclip": __UpperCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] ="altclip" SCREAMING_SNAKE_CASE_ : Optional[int] =True def __init__( self : Any , __A : List[str]=None , __A : List[Any]=None , __A : List[str]=7_6_8 , __A : List[str]=2.6592 , **__A : Dict ): # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __UpperCamelCase = kwargs.pop('text_config_dict' , __A ) __UpperCamelCase = kwargs.pop('vision_config_dict' , __A ) super().__init__(**__A ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __UpperCamelCase = {} # This is the complete result when using `text_config_dict`. __UpperCamelCase = AltCLIPTextConfig(**__A ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `text_config_dict` and `text_config` but with different values. ''' f'''The value `text_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The ''' f'''value `text_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __UpperCamelCase = {} # This is the complete result when using `vision_config_dict`. __UpperCamelCase = AltCLIPVisionConfig(**__A ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __UpperCamelCase = { str(__A ): value for key, value in _vision_config_dict['id2label'].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __UpperCamelCase = ( f'''`{key}` is found in both `vision_config_dict` and `vision_config` but with different ''' f'''values. The value `vision_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: __UpperCamelCase = ( f'''`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. ''' f'''The value `vision_config["{key}"]` will be overriden.''' ) logger.warning(__A ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __UpperCamelCase = {} logger.info('`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.' ) if vision_config is None: __UpperCamelCase = {} logger.info('`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.' ) __UpperCamelCase = AltCLIPTextConfig(**__A ) __UpperCamelCase = AltCLIPVisionConfig(**__A ) __UpperCamelCase = projection_dim __UpperCamelCase = logit_scale_init_value __UpperCamelCase = 1.0 @classmethod def _lowerCamelCase ( cls : Union[str, Any] , __A : AltCLIPTextConfig , __A : AltCLIPVisionConfig , **__A : Optional[Any] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) __UpperCamelCase = self.text_config.to_dict() __UpperCamelCase = self.vision_config.to_dict() __UpperCamelCase = self.__class__.model_type return output
53
0
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class UpperCamelCase__ : '''simple docstring''' def __init__( self , UpperCamelCase__ , UpperCamelCase__=2 , UpperCamelCase__=3 , UpperCamelCase__=4 , UpperCamelCase__=2 , UpperCamelCase__=7 , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=99 , UpperCamelCase__=36 , UpperCamelCase__=2 , UpperCamelCase__=4 , UpperCamelCase__=37 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=512 , UpperCamelCase__=16 , UpperCamelCase__=2 , UpperCamelCase__=0.02 , UpperCamelCase__=6 , UpperCamelCase__=6 , UpperCamelCase__=3 , UpperCamelCase__=4 , UpperCamelCase__=None , UpperCamelCase__=1000 , ) -> str: lowerCamelCase : int = parent lowerCamelCase : Optional[Any] = batch_size lowerCamelCase : Union[str, Any] = num_channels lowerCamelCase : Dict = image_size lowerCamelCase : Union[str, Any] = patch_size lowerCamelCase : Dict = is_training lowerCamelCase : List[str] = use_input_mask lowerCamelCase : int = use_token_type_ids lowerCamelCase : Dict = use_labels lowerCamelCase : Union[str, Any] = vocab_size lowerCamelCase : List[Any] = hidden_size lowerCamelCase : Optional[int] = num_hidden_layers lowerCamelCase : List[str] = num_attention_heads lowerCamelCase : int = intermediate_size lowerCamelCase : List[str] = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : List[Any] = max_position_embeddings lowerCamelCase : Dict = type_vocab_size lowerCamelCase : str = type_sequence_label_size lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = coordinate_size lowerCamelCase : Tuple = shape_size lowerCamelCase : List[Any] = num_labels lowerCamelCase : Tuple = num_choices lowerCamelCase : int = scope lowerCamelCase : Optional[Any] = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) lowerCamelCase : Dict = text_seq_length lowerCamelCase : List[Any] = (image_size // patch_size) ** 2 + 1 lowerCamelCase : Union[str, Any] = self.text_seq_length + self.image_seq_length def _lowercase ( self ) -> int: lowerCamelCase : Optional[int] = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) lowerCamelCase : Dict = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) lowerCamelCase : Optional[int] = bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCamelCase : Dict = bbox[i, j, 3] lowerCamelCase : Optional[int] = bbox[i, j, 1] lowerCamelCase : Any = tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: lowerCamelCase : List[str] = bbox[i, j, 2] lowerCamelCase : Optional[int] = bbox[i, j, 0] lowerCamelCase : Tuple = tmp_coordinate lowerCamelCase : Union[str, Any] = tf.constant(UpperCamelCase__ ) lowerCamelCase : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase : Optional[int] = None if self.use_input_mask: lowerCamelCase : Optional[int] = random_attention_mask([self.batch_size, self.text_seq_length] ) lowerCamelCase : Optional[int] = None if self.use_token_type_ids: lowerCamelCase : List[Any] = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) lowerCamelCase : str = None lowerCamelCase : Optional[int] = None if self.use_labels: lowerCamelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : Tuple = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) lowerCamelCase : Any = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[int]: lowerCamelCase : Optional[int] = TFLayoutLMvaModel(config=UpperCamelCase__ ) # text + image lowerCamelCase : Any = model(UpperCamelCase__ , pixel_values=UpperCamelCase__ , training=UpperCamelCase__ ) lowerCamelCase : Optional[int] = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , training=UpperCamelCase__ , ) lowerCamelCase : Any = model(UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only lowerCamelCase : Dict = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only lowerCamelCase : Union[str, Any] = model({"pixel_values": pixel_values} , training=UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : str = self.num_labels lowerCamelCase : Optional[Any] = TFLayoutLMvaForSequenceClassification(config=UpperCamelCase__ ) lowerCamelCase : str = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , training=UpperCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: lowerCamelCase : int = self.num_labels lowerCamelCase : str = TFLayoutLMvaForTokenClassification(config=UpperCamelCase__ ) lowerCamelCase : Tuple = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , training=UpperCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: lowerCamelCase : Union[str, Any] = 2 lowerCamelCase : Dict = TFLayoutLMvaForQuestionAnswering(config=UpperCamelCase__ ) lowerCamelCase : Tuple = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , start_positions=UpperCamelCase__ , end_positions=UpperCamelCase__ , training=UpperCamelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self ) -> str: lowerCamelCase : List[Any] = self.prepare_config_and_inputs() ((lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase)) : List[str] = config_and_inputs lowerCamelCase : Dict = { "input_ids": input_ids, "bbox": bbox, "pixel_values": pixel_values, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class UpperCamelCase__ (lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Union[str, Any] = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) lowerCamelCase_ : Dict = ( {"""document-question-answering""": TFLayoutLMvaForQuestionAnswering, """feature-extraction""": TFLayoutLMvaModel} if is_tf_available() else {} ) lowerCamelCase_ : int = False lowerCamelCase_ : int = False lowerCamelCase_ : str = False def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: return True def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False ) -> dict: lowerCamelCase : Optional[Any] = copy.deepcopy(UpperCamelCase__ ) if model_class in get_values(UpperCamelCase__ ): lowerCamelCase : List[str] = { k: tf.tile(tf.expand_dims(UpperCamelCase__ , 1 ) , (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(UpperCamelCase__ , tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(UpperCamelCase__ ): lowerCamelCase : Dict = tf.ones(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCamelCase__ ): lowerCamelCase : Dict = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) lowerCamelCase : Optional[Any] = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCamelCase__ ): lowerCamelCase : Dict = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCamelCase__ ): lowerCamelCase : Union[str, Any] = tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=tf.intaa ) return inputs_dict def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : Dict = TFLayoutLMvaModelTester(self ) lowerCamelCase : List[Any] = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def _lowercase ( self ) -> Optional[int]: self.config_tester.run_common_tests() def _lowercase ( self ) -> Optional[int]: lowerCamelCase , lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : List[Any] = model_class(UpperCamelCase__ ) if getattr(UpperCamelCase__ , "hf_compute_loss" , UpperCamelCase__ ): # The number of elements in the loss should be the same as the number of elements in the label lowerCamelCase : Optional[int] = self._prepare_for_class(inputs_dict.copy() , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys() , reverse=UpperCamelCase__ )[0] ] lowerCamelCase : Optional[Any] = added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs lowerCamelCase : int = self._prepare_for_class(inputs_dict.copy() , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase : Dict = prepared_for_class.pop("input_ids" ) lowerCamelCase : Optional[int] = model(UpperCamelCase__ , **UpperCamelCase__ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions lowerCamelCase : List[Any] = self._prepare_for_class(inputs_dict.copy() , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase : Dict = prepared_for_class.pop("input_ids" ) if "labels" in prepared_for_class: lowerCamelCase : List[str] = prepared_for_class["labels"].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: lowerCamelCase : Optional[int] = -100 lowerCamelCase : Any = tf.convert_to_tensor(UpperCamelCase__ ) lowerCamelCase : Dict = model(UpperCamelCase__ , **UpperCamelCase__ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict lowerCamelCase : str = self._prepare_for_class(inputs_dict.copy() , UpperCamelCase__ , return_labels=UpperCamelCase__ ) lowerCamelCase : Optional[int] = model(UpperCamelCase__ )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple lowerCamelCase : Optional[Any] = self._prepare_for_class(inputs_dict.copy() , UpperCamelCase__ , return_labels=UpperCamelCase__ ) # Get keys that were added with the _prepare_for_class function lowerCamelCase : Any = prepared_for_class.keys() - inputs_dict.keys() lowerCamelCase : Any = inspect.signature(model.call ).parameters lowerCamelCase : List[str] = list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple lowerCamelCase : Any = {0: "input_ids"} for label_key in label_keys: lowerCamelCase : Optional[Any] = signature_names.index(UpperCamelCase__ ) lowerCamelCase : int = label_key lowerCamelCase : str = sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple lowerCamelCase : List[str] = [] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: lowerCamelCase : Optional[Any] = prepared_for_class[value] lowerCamelCase : Any = tuple(UpperCamelCase__ ) # Send to model lowerCamelCase : List[Any] = model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def _lowercase ( self ) -> Optional[Any]: ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self ) -> List[str]: ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : List[str] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase : List[str] = type self.model_tester.create_and_check_model(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self ) -> Optional[int]: ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self ) -> Any: ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _lowercase ( self ) -> Union[str, Any]: ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) @slow def _lowercase ( self ) -> Optional[int]: for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase : Dict = TFLayoutLMvaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def A ( ) -> int: lowerCamelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @cached_property def _lowercase ( self ) -> Union[str, Any]: return LayoutLMvaImageProcessor(apply_ocr=UpperCamelCase__ ) if is_vision_available() else None @slow def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Union[str, Any] = TFLayoutLMvaModel.from_pretrained("microsoft/layoutlmv3-base" ) lowerCamelCase : Any = self.default_image_processor lowerCamelCase : Optional[Any] = prepare_img() lowerCamelCase : Dict = image_processor(images=UpperCamelCase__ , return_tensors="tf" ).pixel_values lowerCamelCase : Union[str, Any] = tf.constant([[1, 2]] ) lowerCamelCase : Dict = tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ) , axis=0 ) # forward pass lowerCamelCase : List[Any] = model(input_ids=UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , training=UpperCamelCase__ ) # verify the logits lowerCamelCase : Dict = (1, 199, 768) self.assertEqual(outputs.last_hidden_state.shape , UpperCamelCase__ ) lowerCamelCase : Dict = tf.constant( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCamelCase__ , atol=1e-4 ) )
48
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Any ) -> Optional[Any]: """simple docstring""" with open(__lowercase ) as metadata_file: __UpperCamelCase = json.load(__lowercase ) __UpperCamelCase = LukeConfig(use_entity_aware_attention=__lowercase , **metadata['model_config'] ) # Load in the weights from the checkpoint_path __UpperCamelCase = torch.load(__lowercase , map_location='cpu' ) # Load the entity vocab file __UpperCamelCase = load_entity_vocab(__lowercase ) __UpperCamelCase = RobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks __UpperCamelCase = AddedToken('<ent>' , lstrip=__lowercase , rstrip=__lowercase ) __UpperCamelCase = AddedToken('<ent2>' , lstrip=__lowercase , rstrip=__lowercase ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(__lowercase ) with open(os.path.join(__lowercase , LukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(__lowercase , __lowercase ) __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase ) # Initialize the embeddings of the special tokens __UpperCamelCase = state_dict['embeddings.word_embeddings.weight'] __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['@'] )[0]].unsqueeze(0 ) __UpperCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['#'] )[0]].unsqueeze(0 ) __UpperCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: __UpperCamelCase = F'''encoder.layer.{layer_index}.attention.self.''' __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] __UpperCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks __UpperCamelCase = state_dict['entity_embeddings.entity_embeddings.weight'] __UpperCamelCase = entity_emb[entity_vocab['[MASK]']] __UpperCamelCase = LukeModel(config=__lowercase ).eval() __UpperCamelCase , __UpperCamelCase = model.load_state_dict(__lowercase , strict=__lowercase ) if not (len(__lowercase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F'''Missing keys {', '.join(__lowercase )}. Expected only missing embeddings.position_ids''' ) if not (all(key.startswith('entity_predictions' ) or key.startswith('lm_head' ) for key in unexpected_keys )): raise ValueError( 'Unexpected keys' F''' {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}''' ) # Check outputs __UpperCamelCase = LukeTokenizer.from_pretrained(__lowercase , task='entity_classification' ) __UpperCamelCase = ( 'Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the' ' new world number one avoid a humiliating second- round exit at Wimbledon .' ) __UpperCamelCase = (39, 42) __UpperCamelCase = tokenizer(__lowercase , entity_spans=[span] , add_prefix_space=__lowercase , return_tensors='pt' ) __UpperCamelCase = model(**__lowercase ) # Verify word hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 42, 1024) ) __UpperCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base __UpperCamelCase = torch.Size((1, 42, 768) ) __UpperCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": __UpperCamelCase = torch.Size((1, 1, 1024) ) __UpperCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base __UpperCamelCase = torch.Size((1, 1, 768) ) __UpperCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowercase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(__lowercase ) ) model.save_pretrained(__lowercase ) def lowercase__ ( __lowercase : Dict ) -> List[str]: """simple docstring""" __UpperCamelCase = {} with open(__lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(__lowercase ): __UpperCamelCase , __UpperCamelCase = line.rstrip().split('\t' ) __UpperCamelCase = index return entity_vocab if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) a__ : str =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
53
0
import math def __snake_case ( _UpperCAmelCase = 100 ): __a = sum(i * i for i in range(1 , n + 1 ) ) __a = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(f'{solution() = }')
49
'''simple docstring''' import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class snake_case ( __lowerCamelCase ): """simple docstring""" def _lowerCamelCase ( self : Any ): __UpperCamelCase = tempfile.mkdtemp() __UpperCamelCase = 8 # DPR tok __UpperCamelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase = os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase = dict(zip(__A , range(len(__A ) ) ) ) __UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase = {'unk_token': '<unk>'} __UpperCamelCase = os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(__A , exist_ok=__A ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase = os.path.join(__A , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__A ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__A ) ) def _lowerCamelCase ( self : Tuple ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Optional[int] ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def _lowerCamelCase ( self : Union[str, Any] ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def _lowerCamelCase ( self : str ): shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _lowerCamelCase ( self : Any , __A : bool ): __UpperCamelCase = self.get_dummy_dataset() __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='custom' , ) if from_disk: __UpperCamelCase = os.path.join(self.tmpdirname , 'dataset' ) __UpperCamelCase = os.path.join(self.tmpdirname , 'index.faiss' ) dataset.get_index('embeddings' ).save(os.path.join(self.tmpdirname , 'index.faiss' ) ) dataset.drop_index('embeddings' ) dataset.save_to_disk(os.path.join(self.tmpdirname , 'dataset' ) ) del dataset __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _lowerCamelCase ( self : int ): __UpperCamelCase = Dataset.from_dict( { 'id': ['0', '1'], 'text': ['foo', 'bar'], 'title': ['Foo', 'Bar'], 'embeddings': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('embeddings' , string_factory='Flat' , metric_type=faiss.METRIC_INNER_PRODUCT ) __UpperCamelCase = os.path.join(self.tmpdirname , 'hf_bert_base.hnswSQ8_correct_phi_128.c_index' ) dataset.save_faiss_index('embeddings' , index_file_name + '.index.dpr' ) pickle.dump(dataset['id'] , open(index_file_name + '.index_meta.dpr' , 'wb' ) ) __UpperCamelCase = os.path.join(self.tmpdirname , 'psgs_w100.tsv.pkl' ) __UpperCamelCase = {sample['id']: [sample['text'], sample['title']] for sample in dataset} pickle.dump(__A , open(__A , 'wb' ) ) __UpperCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='legacy' , index_path=self.tmpdirname , ) __UpperCamelCase = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('transformers.models.rag.retrieval_rag.load_dataset' ) as mock_load_dataset: __UpperCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['embeddings', 'id', 'text', 'title'] ) self.assertEqual(len(doc_dicts[0]['id'] ) , __A ) self.assertEqual(doc_dicts[0]['id'][0] , '1' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['id'][0] , '0' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_legacy_index_retriever() __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['text', 'title'] ) self.assertEqual(len(doc_dicts[0]['text'] ) , __A ) self.assertEqual(doc_dicts[0]['text'][0] , 'bar' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['text'][0] , 'foo' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) __UpperCamelCase = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Optional[Any] ): import torch __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_canonical_hf_index_retriever() __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) __UpperCamelCase = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors='pt' , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = ( # noqa: F841 out['context_input_ids'], out['context_attention_mask'], out['retrieved_doc_embeds'], out['doc_ids'], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.get_dpr_ctx_encoder_tokenizer() __UpperCamelCase = 1 __UpperCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) __UpperCamelCase = [[5, 7], [1_0, 1_1]] __UpperCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __UpperCamelCase = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('tokenized_doc_ids', 'tokenized_doc_attention_mask') ) , __A ) # check for doc token related keys in dictionary.
53
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _UpperCAmelCase : Dict = None _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = """▁""" _UpperCAmelCase : Optional[Any] = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } _UpperCAmelCase : List[Any] = { """google/pegasus-xsum""": 5_12, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PegasusTokenizer UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : int , UpperCAmelCase : Optional[int]=None , UpperCAmelCase : List[Any]=None , UpperCAmelCase : Any="<pad>" , UpperCAmelCase : Union[str, Any]="</s>" , UpperCAmelCase : Dict="<unk>" , UpperCAmelCase : Dict="<mask_2>" , UpperCAmelCase : List[str]="<mask_1>" , UpperCAmelCase : Dict=None , UpperCAmelCase : str=103 , **UpperCAmelCase : str , ) -> Any: lowerCamelCase__ : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise TypeError( F"""additional_special_tokens should be of type {type(UpperCAmelCase )}, but is""" F""" {type(UpperCAmelCase )}""" ) lowerCamelCase__ : Optional[Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"""<unk_{i}>""" for i in range(len(UpperCAmelCase ) , self.offset - 1 ) ] if len(set(UpperCAmelCase ) ) != len(UpperCAmelCase ): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' F""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) lowerCamelCase__ : Optional[Any] = additional_special_tokens_extended else: lowerCamelCase__ : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , pad_token=UpperCAmelCase , eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , mask_token=UpperCAmelCase , mask_token_sent=UpperCAmelCase , offset=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : int = vocab_file lowerCamelCase__ : Any = False if not self.vocab_file else True def A_ ( self : Tuple , UpperCAmelCase : str ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( 'There should be 3 special tokens: mask_token, pad_token, and eos_token +' F""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def A_ ( self : List[Any] , UpperCAmelCase : List , UpperCAmelCase : Optional[List] = None , UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return self._special_token_mask(UpperCAmelCase ) elif token_ids_a is None: return self._special_token_mask(UpperCAmelCase ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def A_ ( self : str , UpperCAmelCase : Optional[Any] , UpperCAmelCase : int=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : List[Any] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a__ : List[Any] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys a__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
53
0
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( '''files''' , [ ['''full:README.md''', '''dataset_infos.json'''], ['''empty:README.md''', '''dataset_infos.json'''], ['''dataset_infos.json'''], ['''full:README.md'''], ] , ) def A (__A : Tuple , __A : Dict ) -> Any: """simple docstring""" UpperCAmelCase_ = tmp_path_factory.mktemp('''dset_infos_dir''' ) if "full:README.md" in files: with open(dataset_infos_dir / '''README.md''' , '''w''' ) as f: f.write('''---\ndataset_info:\n dataset_size: 42\n---''' ) if "empty:README.md" in files: with open(dataset_infos_dir / '''README.md''' , '''w''' ) as f: f.write('''''' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / '''dataset_infos.json''' , '''w''' ) as f: f.write('''{"default": {"dataset_size": 42}}''' ) UpperCAmelCase_ = DatasetInfosDict.from_directory(__A ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( '''dataset_info''' , [ DatasetInfo(), DatasetInfo( description='''foo''' , features=Features({'''a''': Value('''int32''' )} ) , builder_name='''builder''' , config_name='''config''' , version='''1.0.0''' , splits=[{'''name''': '''train'''}] , download_size=42 , ), ] , ) def A (__A : Optional[int] , __A : DatasetInfo ) -> Tuple: """simple docstring""" UpperCAmelCase_ = str(__A ) dataset_info.write_to_directory(__A ) UpperCAmelCase_ = DatasetInfo.from_directory(__A ) assert dataset_info == reloaded assert os.path.exists(os.path.join(__A , '''dataset_info.json''' ) ) def A () -> Optional[int]: """simple docstring""" UpperCAmelCase_ = DatasetInfo( description='''foo''' , citation='''bar''' , homepage='''https://foo.bar''' , license='''CC0''' , features=Features({'''a''': Value('''int32''' )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name='''builder''' , config_name='''config''' , version='''1.0.0''' , splits=[{'''name''': '''train''', '''num_examples''': 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) UpperCAmelCase_ = dataset_info._to_yaml_dict() assert sorted(__A ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) UpperCAmelCase_ = yaml.safe_dump(__A ) UpperCAmelCase_ = yaml.safe_load(__A ) assert dataset_info_yaml_dict == reloaded def A () -> Any: """simple docstring""" UpperCAmelCase_ = DatasetInfo() UpperCAmelCase_ = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( '''dataset_infos_dict''' , [ DatasetInfosDict(), DatasetInfosDict({'''default''': DatasetInfo()} ), DatasetInfosDict({'''my_config_name''': DatasetInfo()} ), DatasetInfosDict( { '''default''': DatasetInfo( description='''foo''' , features=Features({'''a''': Value('''int32''' )} ) , builder_name='''builder''' , config_name='''config''' , version='''1.0.0''' , splits=[{'''name''': '''train'''}] , download_size=42 , ) } ), DatasetInfosDict( { '''v1''': DatasetInfo(dataset_size=42 ), '''v2''': DatasetInfo(dataset_size=1337 ), } ), ] , ) def A (__A : Optional[Any] , __A : DatasetInfosDict ) -> List[Any]: """simple docstring""" UpperCAmelCase_ = str(__A ) dataset_infos_dict.write_to_directory(__A ) UpperCAmelCase_ = DatasetInfosDict.from_directory(__A ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): UpperCAmelCase_ = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml UpperCAmelCase_ = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(__A , '''README.md''' ) )
51
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any]=False ) -> Tuple: """simple docstring""" try: __UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCamelCase = default else: # KEY is set, convert it to True or False. try: __UpperCamelCase = strtobool(__lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value a__ : str =parse_flag_from_env('''RUN_SLOW''', default=False) a__ : Union[str, Any] =parse_flag_from_env('''RUN_REMOTE''', default=False) a__ : List[str] =parse_flag_from_env('''RUN_LOCAL''', default=True) a__ : Optional[int] =parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression a__ : Any =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') a__ : Optional[int] =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') a__ : List[str] =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio a__ : Any =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam a__ : Tuple =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility a__ : Union[str, Any] =pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows a__ : int =pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase__ ( __lowercase : Optional[Any] ) -> Optional[int]: """simple docstring""" try: import faiss # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires faiss' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Any: """simple docstring""" try: import regex # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires regex' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Tuple ) -> List[Any]: """simple docstring""" try: import elasticsearch # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires elasticsearch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" try: import sqlalchemy # noqa except ImportError: __UpperCamelCase = unittest.skip('test requires sqlalchemy' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[str] ) -> List[str]: """simple docstring""" if not config.TORCH_AVAILABLE: __UpperCamelCase = unittest.skip('test requires PyTorch' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[Any] ) -> List[str]: """simple docstring""" if not config.TF_AVAILABLE: __UpperCamelCase = unittest.skip('test requires TensorFlow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : int ) -> Union[str, Any]: """simple docstring""" if not config.JAX_AVAILABLE: __UpperCamelCase = unittest.skip('test requires JAX' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> Optional[Any]: """simple docstring""" if not config.PIL_AVAILABLE: __UpperCamelCase = unittest.skip('test requires Pillow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('test requires transformers' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : int ) -> int: """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('test requires tiktoken' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : str ) -> Any: """simple docstring""" def _require_spacy_model(__lowercase : Any ): try: import spacy # noqa F401 spacy.load(__lowercase ) except ImportError: return unittest.skip('test requires spacy' )(__lowercase ) except OSError: return unittest.skip('test requires spacy model \'{}\''.format(__lowercase ) )(__lowercase ) else: return test_case return _require_spacy_model def lowercase__ ( __lowercase : Union[str, Any] ) -> str: """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('test requires pyspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('test requires joblibspark' )(__lowercase ) else: return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: __UpperCamelCase = unittest.skip('test is slow' )(__lowercase ) return test_case def lowercase__ ( __lowercase : List[Any] ) -> List[str]: """simple docstring""" if not _run_local_tests or _run_local_tests == 0: __UpperCamelCase = unittest.skip('test is local' )(__lowercase ) return test_case def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCamelCase = unittest.skip('test is packaged' )(__lowercase ) return test_case def lowercase__ ( __lowercase : Optional[int] ) -> Any: """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: __UpperCamelCase = unittest.skip('test requires remote' )(__lowercase ) return test_case def lowercase__ ( *__lowercase : Optional[Any] ) -> Tuple: """simple docstring""" def decorate(cls : int ): for name, fn in cls.__dict__.items(): if callable(__lowercase ) and name.startswith('test' ): for decorator in decorators: __UpperCamelCase = decorator(__lowercase ) setattr(cls , __lowercase , __lowercase ) return cls return decorate class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =0 SCREAMING_SNAKE_CASE_ : List[Any] =1 SCREAMING_SNAKE_CASE_ : Union[str, Any] =2 @contextmanager def lowercase__ ( __lowercase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , __lowercase : Dict=1e-16 ) -> List[Any]: """simple docstring""" __UpperCamelCase = requests.Session().request def timeout_request(__lowercase : List[Any] , __lowercase : Tuple , __lowercase : List[Any] , **__lowercase : List[str] ): # Change the url to an invalid url so that the connection hangs __UpperCamelCase = 'https://10.255.255.1' if kwargs.get('timeout' ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCamelCase = timeout try: return online_request(__lowercase , __lowercase , **__lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCamelCase = url __UpperCamelCase = e.args[0] __UpperCamelCase = (max_retry_error.args[0].replace('10.255.255.1' , F'''OfflineMock[{url}]''' ),) __UpperCamelCase = (max_retry_error,) raise def raise_connection_error(__lowercase : int , __lowercase : List[str] , **__lowercase : Union[str, Any] ): raise requests.ConnectionError('Offline mode is enabled.' , request=__lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('requests.Session.send' , __lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('requests.Session.request' , __lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('datasets.config.HF_DATASETS_OFFLINE' , __lowercase ): yield else: raise ValueError('Please use a value from the OfflineSimulationMode enum.' ) @contextmanager def lowercase__ ( *__lowercase : Any , **__lowercase : Dict ) -> Dict: """simple docstring""" __UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__lowercase , **__lowercase ) as tmp_dir: try: os.chdir(__lowercase ) yield finally: os.chdir(__lowercase ) @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase__ ( ) -> Optional[Any]: """simple docstring""" import gc gc.collect() __UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase__ ( __lowercase : List[str] , __lowercase : int ) -> Union[str, Any]: """simple docstring""" return deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() == deepcopy(__lowercase ).integers(0 , 100 , 10 ).tolist() def lowercase__ ( __lowercase : str ) -> List[str]: """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(__lowercase : List[Any] , *__lowercase : Tuple , **__lowercase : Union[str, Any] ): try: return func(*__lowercase , **__lowercase ) except HTTPError as err: if str(__lowercase ).startswith('500' ) or str(__lowercase ).startswith('502' ): pytest.xfail(str(__lowercase ) ) raise err return decorator.decorator(_wrapper , __lowercase ) class snake_case : """simple docstring""" def __init__( self : int , __A : Any , __A : str , __A : List[Any] ): __UpperCamelCase = returncode __UpperCamelCase = stdout __UpperCamelCase = stderr async def lowercase__ ( __lowercase : Any , __lowercase : Optional[int] ) -> str: """simple docstring""" while True: __UpperCamelCase = await stream.readline() if line: callback(__lowercase ) else: break async def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any]=None , __lowercase : Any=None , __lowercase : Optional[Any]=None , __lowercase : int=False , __lowercase : List[Any]=False ) -> _RunOutput: """simple docstring""" if echo: print('\nRunning: ' , ' '.join(__lowercase ) ) __UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=__lowercase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=__lowercase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCamelCase = [] __UpperCamelCase = [] def tee(__lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : Tuple="" ): __UpperCamelCase = line.decode('utf-8' ).rstrip() sink.append(__lowercase ) if not quiet: print(__lowercase , __lowercase , file=__lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda __lowercase : tee(__lowercase , __lowercase , sys.stdout , label='stdout:' ) ), _read_stream(p.stderr , lambda __lowercase : tee(__lowercase , __lowercase , sys.stderr , label='stderr:' ) ), ] , timeout=__lowercase , ) return _RunOutput(await p.wait() , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict , __lowercase : Any=None , __lowercase : int=None , __lowercase : int=180 , __lowercase : int=False , __lowercase : str=True ) -> _RunOutput: """simple docstring""" __UpperCamelCase = asyncio.get_event_loop() __UpperCamelCase = loop.run_until_complete( _stream_subprocess(__lowercase , env=__lowercase , stdin=__lowercase , timeout=__lowercase , quiet=__lowercase , echo=__lowercase ) ) __UpperCamelCase = ' '.join(__lowercase ) if result.returncode > 0: __UpperCamelCase = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def lowercase__ ( ) -> List[str]: """simple docstring""" __UpperCamelCase = os.environ.get('PYTEST_XDIST_WORKER' , 'gw0' ) __UpperCamelCase = re.sub(R'^gw' , '' , __lowercase , 0 , re.M ) return int(__lowercase ) def lowercase__ ( ) -> List[Any]: """simple docstring""" __UpperCamelCase = 29500 __UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
53
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : Optional[int] = logging.get_logger(__name__) __lowerCamelCase : str = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/config.json""" # See all FNet models at https://huggingface.co/models?filter=fnet } class A__ ( __snake_case ): _UpperCAmelCase :str = 'fnet' def __init__( self , A_=3_2000 , A_=768 , A_=12 , A_=3072 , A_="gelu_new" , A_=0.1 , A_=512 , A_=4 , A_=0.02 , A_=1e-12 , A_=False , A_=512 , A_=3 , A_=1 , A_=2 , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : Any = vocab_size UpperCamelCase : Tuple = max_position_embeddings UpperCamelCase : Any = hidden_size UpperCamelCase : List[str] = num_hidden_layers UpperCamelCase : Any = intermediate_size UpperCamelCase : Optional[Any] = hidden_act UpperCamelCase : int = hidden_dropout_prob UpperCamelCase : Dict = initializer_range UpperCamelCase : Union[str, Any] = type_vocab_size UpperCamelCase : str = layer_norm_eps UpperCamelCase : List[str] = use_tpu_fourier_optimizations UpperCamelCase : Optional[int] = tpu_short_seq_length
52
'''simple docstring''' # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys a__ : Tuple ='''3''' print('''Python version:''', sys.version) print('''OS platform:''', platform.platform()) print('''OS architecture:''', platform.machine()) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) except ImportError: print('''Torch version:''', None) try: import transformers print('''transformers version:''', transformers.__version__) except ImportError: print('''transformers version:''', None)
53
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a__ : Optional[Any] = logging.get_logger(__name__) a__ : Optional[Any] = '''▁''' a__ : Tuple = {'''vocab_file''': '''spiece.model'''} a__ : Tuple = { '''vocab_file''': { '''google/reformer-crime-and-punishment''': ( '''https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model''' ) } } a__ : Union[str, Any] = { '''google/reformer-crime-and-punishment''': 5_2_4_2_8_8, } class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Tuple = VOCAB_FILES_NAMES snake_case__ : Tuple = PRETRAINED_VOCAB_FILES_MAP snake_case__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ : Optional[Any] = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Tuple="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : List[Any]=[] , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : Optional[int] , ) -> None: __SCREAMING_SNAKE_CASE = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = vocab_file __SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) @property def UpperCAmelCase_ ( self : Dict ) -> str: return self.sp_model.get_piece_size() def UpperCAmelCase_ ( self : List[Any] ) -> Dict[str, int]: __SCREAMING_SNAKE_CASE = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.__dict__.copy() __SCREAMING_SNAKE_CASE = None return state def __setstate__( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : str ) -> List[str]: return self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : Union[str, Any] ) -> Optional[int]: return self.sp_model.piece_to_id(UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : Any ) -> List[str]: if index < self.sp_model.get_piece_size(): __SCREAMING_SNAKE_CASE = self.sp_model.IdToPiece(UpperCAmelCase__ ) return token def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : List[Any] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCAmelCase__ ) + token __SCREAMING_SNAKE_CASE = [] else: current_sub_tokens.append(UpperCAmelCase__ ) out_string += self.sp_model.decode(UpperCAmelCase__ ) return out_string.strip() def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , "wb" ) as fi: __SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,)
54
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Optional[int] , __lowercase : Tuple , __lowercase : Tuple ) -> Tuple: """simple docstring""" return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def lowercase__ ( __lowercase : Optional[int] , __lowercase : Dict , __lowercase : List[str] , __lowercase : List[str]="attention" ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) __UpperCamelCase = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) __UpperCamelCase = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) __UpperCamelCase = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __UpperCamelCase = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) __UpperCamelCase = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowercase__ ( __lowercase : Tuple , __lowercase : Dict , __lowercase : int , __lowercase : List[Any]=False ) -> Optional[Any]: """simple docstring""" if split_mlp_wi: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] __UpperCamelCase = (wi_a, wi_a) else: __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] __UpperCamelCase = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : List[str] , __lowercase : Optional[int] ) -> str: """simple docstring""" return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def lowercase__ ( __lowercase : dict , *, __lowercase : int , __lowercase : bool , __lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = traverse_util.flatten_dict(variables['target'] ) __UpperCamelCase = {'/'.join(__lowercase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __UpperCamelCase = 'encoder/encoder/mlp/wi_0/kernel' in old print('Split MLP:' , __lowercase ) __UpperCamelCase = collections.OrderedDict() # Shared embeddings. __UpperCamelCase = old['token_embedder/embedding'] # Encoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'encoder' , 'attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'encoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'encoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , __lowercase , 'encoder' ).T __UpperCamelCase = old['encoder/encoder_norm/scale'] if not scalable_attention: __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'encoder' ).T __UpperCamelCase = tax_relpos_bias_lookup( __lowercase , 0 , 'decoder' ).T if not is_encoder_only: # Decoder. for i in range(__lowercase ): # Block i, layer 0 (Self Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_self_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'self_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 1 (Cross Attention). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_cross_attention_layer_norm' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = tax_attention_lookup(__lowercase , __lowercase , 'decoder' , 'encoder_decoder_attention' ) __UpperCamelCase = layer_norm __UpperCamelCase = k.T __UpperCamelCase = o.T __UpperCamelCase = q.T __UpperCamelCase = v.T # Block i, layer 2 (MLP). __UpperCamelCase = tax_layer_norm_lookup(__lowercase , __lowercase , 'decoder' , 'pre_mlp_layer_norm' ) __UpperCamelCase , __UpperCamelCase = tax_mlp_lookup(__lowercase , __lowercase , 'decoder' , __lowercase ) __UpperCamelCase = layer_norm if split_mlp_wi: __UpperCamelCase = wi[0].T __UpperCamelCase = wi[1].T else: __UpperCamelCase = wi.T __UpperCamelCase = wo.T if scalable_attention: # convert the rel_embedding of each layer __UpperCamelCase = tax_relpos_bias_lookup(__lowercase , __lowercase , 'decoder' ).T __UpperCamelCase = old['decoder/decoder_norm/scale'] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __UpperCamelCase = old['decoder/logits_dense/kernel'].T return new def lowercase__ ( __lowercase : Optional[Any] , __lowercase : bool ) -> int: """simple docstring""" __UpperCamelCase = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __UpperCamelCase = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) __UpperCamelCase = state_dict['shared.weight'] return state_dict def lowercase__ ( __lowercase : List[str] , __lowercase : Dict , __lowercase : str , __lowercase : int , __lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = checkpoints.load_tax_checkpoint(__lowercase ) __UpperCamelCase = convert_tax_to_pytorch( __lowercase , num_layers=config.num_layers , is_encoder_only=__lowercase , scalable_attention=__lowercase ) __UpperCamelCase = make_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase , strict=__lowercase ) def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : Dict , __lowercase : List[str] , __lowercase : bool = False , __lowercase : bool = False , ) -> Optional[int]: """simple docstring""" __UpperCamelCase = MTaConfig.from_json_file(__lowercase ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __UpperCamelCase = UMTaEncoderModel(__lowercase ) else: __UpperCamelCase = UMTaForConditionalGeneration(__lowercase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowercase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowercase ) print('Done' ) if __name__ == "__main__": a__ : List[Any] =argparse.ArgumentParser(description='''Converts a native T5X checkpoint into a PyTorch checkpoint.''') # Required parameters parser.add_argument( '''--t5x_checkpoint_path''', default=None, type=str, required=True, help='''Path to the T5X checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--is_encoder_only''', action='''store_true''', help='''Check if the model is encoder-decoder model''', default=False ) parser.add_argument( '''--scalable_attention''', action='''store_true''', help='''Whether the model uses scaled attention (umt5 model)''', default=False, ) a__ : List[str] =parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
53
0
'''simple docstring''' from __future__ import annotations from fractions import Fraction def __snake_case ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __snake_case ( UpperCAmelCase_ : int ): lowerCamelCase_ = [] lowerCamelCase_ = 11 lowerCamelCase_ = int("1" + "0" * digit_len ) for num in range(UpperCAmelCase_ , UpperCAmelCase_ ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(UpperCAmelCase_ , UpperCAmelCase_ ): solutions.append(F'''{num}/{den}''' ) den += 1 num += 1 lowerCamelCase_ = 10 return solutions def __snake_case ( UpperCAmelCase_ : int = 2 ): lowerCamelCase_ = 1.0 for fraction in fraction_list(UpperCAmelCase_ ): lowerCamelCase_ = Fraction(UpperCAmelCase_ ) result *= frac.denominator / frac.numerator return int(UpperCAmelCase_ ) if __name__ == "__main__": print(solution())
55
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : List[Any] ="BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Optional[int] =("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , __A : Optional[int] , __A : List[Any] ): __UpperCamelCase = False super().__init__(__A , __A ) __UpperCamelCase = self.image_processor def __call__( self : List[Any] , __A : ImageInput = None , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __UpperCamelCase = self.tokenizer __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding # add pixel_values __UpperCamelCase = self.image_processor(__A , return_tensors=__A ) if text is not None: __UpperCamelCase = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) else: __UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def _lowerCamelCase ( self : List[Any] , *__A : Dict , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowerCamelCase ( self : List[Any] , *__A : List[str] , **__A : Dict ): return self.tokenizer.decode(*__A , **__A ) @property def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
0
'''simple docstring''' import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants a : Any = Mapping[str, np.ndarray] a : Tuple = Mapping[str, Any] # Is a nested dict. a : Union[str, Any] = 0.01 @dataclasses.dataclass(frozen=_lowerCamelCase ) class a : snake_case_ = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. snake_case_ = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. snake_case_ = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. snake_case_ = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. snake_case_ = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions snake_case_ = None # Optional remark about the protein. Included as a comment in output PDB # files snake_case_ = None # Templates used to generate this protein (prediction-only) snake_case_ = None # Chain corresponding to each parent snake_case_ = None def __magic_name__ ( __UpperCAmelCase ) -> Protein: '''simple docstring''' snake_case_ = r'''(\[[A-Z]+\]\n)''' snake_case_ = [tag.strip() for tag in re.split(__UpperCAmelCase, __UpperCAmelCase ) if len(__UpperCAmelCase ) > 0] snake_case_ = zip(tags[0::2], [l.split('''\n''' ) for l in tags[1::2]] ) snake_case_ = ["N", "CA", "C"] snake_case_ = None snake_case_ = None snake_case_ = None for g in groups: if "[PRIMARY]" == g[0]: snake_case_ = g[1][0].strip() for i in range(len(__UpperCAmelCase ) ): if seq[i] not in residue_constants.restypes: snake_case_ = '''X''' # FIXME: strings are immutable snake_case_ = np.array( [residue_constants.restype_order.get(__UpperCAmelCase, residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: snake_case_ = [] for axis in range(3 ): tertiary.append(list(map(__UpperCAmelCase, g[1][axis].split() ) ) ) snake_case_ = np.array(__UpperCAmelCase ) snake_case_ = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(__UpperCAmelCase ): snake_case_ = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: snake_case_ = np.array(list(map({'''-''': 0, '''+''': 1}.get, g[1][0].strip() ) ) ) snake_case_ = np.zeros( ( len(__UpperCAmelCase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(__UpperCAmelCase ): snake_case_ = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=__UpperCAmelCase, atom_mask=__UpperCAmelCase, aatype=__UpperCAmelCase, residue_index=np.arange(len(__UpperCAmelCase ) ), b_factors=__UpperCAmelCase, ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase = 0 ) -> List[str]: '''simple docstring''' snake_case_ = [] snake_case_ = prot.remark if remark is not None: pdb_headers.append(F"REMARK {remark}" ) snake_case_ = prot.parents snake_case_ = prot.parents_chain_index if parents is not None and parents_chain_index is not None: snake_case_ = [p for i, p in zip(__UpperCAmelCase, __UpperCAmelCase ) if i == chain_id] if parents is None or len(__UpperCAmelCase ) == 0: snake_case_ = ['''N/A'''] pdb_headers.append(F"PARENT {' '.join(__UpperCAmelCase )}" ) return pdb_headers def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> str: '''simple docstring''' snake_case_ = [] snake_case_ = pdb_str.split('''\n''' ) snake_case_ = prot.remark if remark is not None: out_pdb_lines.append(F"REMARK {remark}" ) snake_case_ = 42 if prot.parents is not None and len(prot.parents ) > 0: snake_case_ = [] if prot.parents_chain_index is not None: snake_case_ = {} for p, i in zip(prot.parents, prot.parents_chain_index ): parent_dict.setdefault(str(__UpperCAmelCase ), [] ) parent_dict[str(__UpperCAmelCase )].append(__UpperCAmelCase ) snake_case_ = max([int(__UpperCAmelCase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): snake_case_ = parent_dict.get(str(__UpperCAmelCase ), ['''N/A'''] ) parents_per_chain.append(__UpperCAmelCase ) else: parents_per_chain.append(list(prot.parents ) ) else: snake_case_ = [['''N/A''']] def make_parent_line(__UpperCAmelCase ) -> str: return F"PARENT {' '.join(__UpperCAmelCase )}" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) snake_case_ = 0 for i, l in enumerate(__UpperCAmelCase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(__UpperCAmelCase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(__UpperCAmelCase ): snake_case_ = parents_per_chain[chain_counter] else: snake_case_ = ['''N/A'''] out_pdb_lines.append(make_parent_line(__UpperCAmelCase ) ) return "\n".join(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> str: '''simple docstring''' snake_case_ = residue_constants.restypes + ['''X'''] def res_atoa(__UpperCAmelCase ) -> str: return residue_constants.restype_atoa.get(restypes[r], '''UNK''' ) snake_case_ = residue_constants.atom_types snake_case_ = [] snake_case_ = prot.atom_mask snake_case_ = prot.aatype snake_case_ = prot.atom_positions snake_case_ = prot.residue_index.astype(np.intaa ) snake_case_ = prot.b_factors snake_case_ = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError('''Invalid aatypes.''' ) snake_case_ = get_pdb_headers(__UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: pdb_lines.extend(__UpperCAmelCase ) snake_case_ = aatype.shape[0] snake_case_ = 1 snake_case_ = 0 snake_case_ = string.ascii_uppercase snake_case_ = None # Add all atom sites. for i in range(__UpperCAmelCase ): snake_case_ = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(__UpperCAmelCase, atom_positions[i], atom_mask[i], b_factors[i] ): if mask < 0.5: continue snake_case_ = '''ATOM''' snake_case_ = atom_name if len(__UpperCAmelCase ) == 4 else F" {atom_name}" snake_case_ = '''''' snake_case_ = '''''' snake_case_ = 1.0_0 snake_case_ = atom_name[0] # Protein supports only C, N, O, S, this works. snake_case_ = '''''' snake_case_ = '''A''' if chain_index is not None: snake_case_ = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! snake_case_ = ( F"{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}" F"{res_name_a:>3} {chain_tag:>1}" F"{residue_index[i]:>4}{insertion_code:>1} " F"{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}" F"{occupancy:>6.2f}{b_factor:>6.2f} " F"{element:>2}{charge:>2}" ) pdb_lines.append(__UpperCAmelCase ) atom_index += 1 snake_case_ = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: snake_case_ = True snake_case_ = chain_index[i + 1] if should_terminate: # Close the chain. snake_case_ = '''TER''' snake_case_ = ( F"{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}" ) pdb_lines.append(__UpperCAmelCase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(__UpperCAmelCase, __UpperCAmelCase ) ) pdb_lines.append('''END''' ) pdb_lines.append('''''' ) return "\n".join(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase = None, __UpperCAmelCase = None, __UpperCAmelCase = None, __UpperCAmelCase = None, __UpperCAmelCase = None, ) -> Protein: '''simple docstring''' return Protein( aatype=features['''aatype'''], atom_positions=result['''final_atom_positions'''], atom_mask=result['''final_atom_mask'''], residue_index=features['''residue_index'''] + 1, b_factors=b_factors if b_factors is not None else np.zeros_like(result['''final_atom_mask'''] ), chain_index=__UpperCAmelCase, remark=__UpperCAmelCase, parents=__UpperCAmelCase, parents_chain_index=__UpperCAmelCase, )
56
'''simple docstring''' from __future__ import annotations from typing import Any class snake_case ( __lowerCamelCase ): """simple docstring""" pass class snake_case : """simple docstring""" def __init__( self : List[Any] , __A : Any ): __UpperCamelCase = data __UpperCamelCase = None def __iter__( self : Optional[Any] ): __UpperCamelCase = self __UpperCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(__A ) yield node.data __UpperCamelCase = node.next_node @property def _lowerCamelCase ( self : List[str] ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": a__ : Dict =Node(1) a__ : Optional[int] =Node(2) a__ : List[str] =Node(3) a__ : Optional[int] =Node(4) print(root_node.has_loop) # False a__ : str =root_node.next_node print(root_node.has_loop) # True a__ : Optional[int] =Node(5) a__ : List[Any] =Node(6) a__ : int =Node(5) a__ : Tuple =Node(6) print(root_node.has_loop) # False a__ : str =Node(1) print(root_node.has_loop) # False
53
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A : str = logging.get_logger(__name__) A : Union[str, Any] = { "microsoft/swin-tiny-patch4-window7-224": ( "https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json" ), # See all Swin models at https://huggingface.co/models?filter=swin } class _UpperCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : Optional[int] ="""swin""" __UpperCAmelCase : Any ={ """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , __a=2_24 , __a=4 , __a=3 , __a=96 , __a=[2, 2, 6, 2] , __a=[3, 6, 12, 24] , __a=7 , __a=4.0 , __a=True , __a=0.0 , __a=0.0 , __a=0.1 , __a="gelu" , __a=False , __a=0.0_2 , __a=1e-5 , __a=32 , __a=None , __a=None , **__a , ): super().__init__(**__a ) __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = embed_dim __lowerCAmelCase = depths __lowerCAmelCase = len(__a ) __lowerCAmelCase = num_heads __lowerCAmelCase = window_size __lowerCAmelCase = mlp_ratio __lowerCAmelCase = qkv_bias __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = drop_path_rate __lowerCAmelCase = hidden_act __lowerCAmelCase = use_absolute_embeddings __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = initializer_range __lowerCAmelCase = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowerCAmelCase = int(embed_dim * 2 ** (len(__a ) - 1) ) __lowerCAmelCase = ["stem"] + [f"stage{idx}" for idx in range(1 , len(__a ) + 1 )] __lowerCAmelCase , __lowerCAmelCase = get_aligned_output_features_output_indices( out_features=__a , out_indices=__a , stage_names=self.stage_names ) class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : Optional[int] =version.parse("""1.11""" ) @property def snake_case ( self ): return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def snake_case ( self ): return 1e-4
57
'''simple docstring''' a__ : Optional[Any] =256 # Modulus to hash a string a__ : Dict =1_000_003 def lowercase__ ( __lowercase : str , __lowercase : str ) -> bool: """simple docstring""" __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) if p_len > t_len: return False __UpperCamelCase = 0 __UpperCamelCase = 0 __UpperCamelCase = 1 # Calculating the hash of pattern and substring of text for i in range(__lowercase ): __UpperCamelCase = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus __UpperCamelCase = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue __UpperCamelCase = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash __UpperCamelCase = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: """simple docstring""" __UpperCamelCase = 'abc1abc12' __UpperCamelCase = 'alskfjaldsabc1abc1abc12k23adsfabcabc' __UpperCamelCase = 'alskfjaldsk23adsfabcabc' assert rabin_karp(__lowercase , __lowercase ) and not rabin_karp(__lowercase , __lowercase ) # Test 2) __UpperCamelCase = 'ABABX' __UpperCamelCase = 'ABABZABABYABABX' assert rabin_karp(__lowercase , __lowercase ) # Test 3) __UpperCamelCase = 'AAAB' __UpperCamelCase = 'ABAAAAAB' assert rabin_karp(__lowercase , __lowercase ) # Test 4) __UpperCamelCase = 'abcdabcy' __UpperCamelCase = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(__lowercase , __lowercase ) # Test 5) __UpperCamelCase = 'Lü' __UpperCamelCase = 'Lüsai' assert rabin_karp(__lowercase , __lowercase ) __UpperCamelCase = 'Lue' assert not rabin_karp(__lowercase , __lowercase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
0
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class a_ ( snake_case_ ): '''simple docstring''' UpperCamelCase = 42 UpperCamelCase = 42 def __init__( self , A , A ) -> Optional[Any]: super().__init__() self.register_modules(unet=A , scheduler=A ) @torch.no_grad() def __call__( self , A = 1 , A = 2000 , A = None , A = "pil" , A = True , **A , ) -> Union[ImagePipelineOutput, Tuple]: _SCREAMING_SNAKE_CASE = self.unet.config.sample_size _SCREAMING_SNAKE_CASE = (batch_size, 3, img_size, img_size) _SCREAMING_SNAKE_CASE = self.unet _SCREAMING_SNAKE_CASE = randn_tensor(A , generator=A ) * self.scheduler.init_noise_sigma _SCREAMING_SNAKE_CASE = sample.to(self.device ) self.scheduler.set_timesteps(A ) self.scheduler.set_sigmas(A ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): _SCREAMING_SNAKE_CASE = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): _SCREAMING_SNAKE_CASE = self.unet(A , A ).sample _SCREAMING_SNAKE_CASE = self.scheduler.step_correct(A , A , generator=A ).prev_sample # prediction step _SCREAMING_SNAKE_CASE = model(A , A ).sample _SCREAMING_SNAKE_CASE = self.scheduler.step_pred(A , A , A , generator=A ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = output.prev_sample, output.prev_sample_mean _SCREAMING_SNAKE_CASE = sample_mean.clamp(0 , 1 ) _SCREAMING_SNAKE_CASE = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _SCREAMING_SNAKE_CASE = self.numpy_to_pil(A ) if not return_dict: return (sample,) return ImagePipelineOutput(images=A )
58
'''simple docstring''' from __future__ import annotations class snake_case : """simple docstring""" def __init__( self : Optional[int] , __A : list[list[int]] ): __UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__A ) != 0: __UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__A ) != cols: raise error for value in row: if not isinstance(__A , (int, float) ): raise error __UpperCamelCase = rows else: __UpperCamelCase = [] def _lowerCamelCase ( self : int ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def _lowerCamelCase ( self : str ): return len(self.rows ) @property def _lowerCamelCase ( self : Any ): return len(self.rows[0] ) @property def _lowerCamelCase ( self : Optional[Any] ): return (self.num_rows, self.num_columns) @property def _lowerCamelCase ( self : Dict ): return self.order[0] == self.order[1] def _lowerCamelCase ( self : Any ): __UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Any ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def _lowerCamelCase ( self : List[str] ): return bool(self.determinant() ) def _lowerCamelCase ( self : Dict , __A : int , __A : int ): __UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__A ).determinant() def _lowerCamelCase ( self : Dict , __A : int , __A : int ): if (row + column) % 2 == 0: return self.get_minor(__A , __A ) return -1 * self.get_minor(__A , __A ) def _lowerCamelCase ( self : List[str] ): return Matrix( [ [self.get_minor(__A , __A ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def _lowerCamelCase ( self : Union[str, Any] ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__A ) def _lowerCamelCase ( self : Dict ): __UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Optional[Any] ): return str(self.rows ) def __str__( self : Union[str, Any] ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__A ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def _lowerCamelCase ( self : List[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in row: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__A ) else: __UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def _lowerCamelCase ( self : Optional[Any] , __A : list[int] , __A : int | None = None ): __UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__A , __A ): raise type_error for value in column: if not isinstance(__A , (int, float) ): raise type_error if len(__A ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: __UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: __UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Tuple , __A : object ): if not isinstance(__A , __A ): return NotImplemented return self.rows == other.rows def __ne__( self : Any , __A : object ): return not self == other def __neg__( self : List[Any] ): return self * -1 def __add__( self : List[str] , __A : Matrix ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : str , __A : Matrix ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : str , __A : Matrix | int | float ): if isinstance(__A , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__A , __A ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__A , __A ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Union[str, Any] , __A : int ): if not isinstance(__A , __A ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) __UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def _lowerCamelCase ( cls : Tuple , __A : list[int] , __A : list[int] ): return sum(row[i] * column[i] for i in range(len(__A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
53
0
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() __lowerCamelCase = logging.get_logger(__name__) __lowerCamelCase = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn.grep_linear""": """encoder.layers.*.attention.gru_rel_pos_linear""", """self_attn.relative_attention_bias""": """encoder.layers.*.attention.rel_attn_embed""", """self_attn.grep_a""": """encoder.layers.*.attention.gru_rel_pos_const""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } __lowerCamelCase = [ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def UpperCamelCase ( __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : int ): for attribute in key.split("." ): snake_case : int = getattr(__lowerCamelCase , __lowerCamelCase ) if weight_type is not None: snake_case : Union[str, Any] = getattr(__lowerCamelCase , __lowerCamelCase ).shape else: snake_case : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": snake_case : Optional[Any] = value elif weight_type == "weight_g": snake_case : str = value elif weight_type == "weight_v": snake_case : Dict = value elif weight_type == "bias": snake_case : Dict = value else: snake_case : Tuple = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def UpperCamelCase ( __lowerCamelCase : int , __lowerCamelCase : Any ): snake_case : List[Any] = [] snake_case : List[Any] = fairseq_model.state_dict() snake_case : str = hf_model.feature_extractor for name, value in fairseq_dict.items(): snake_case : Tuple = False if "conv_layers" in name: load_conv_layer( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , hf_model.config.feat_extract_norm == "group" , ) snake_case : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: snake_case : Dict = True if "*" in mapped_key: snake_case : Union[str, Any] = name.split(__lowerCamelCase )[0].split("." )[-2] snake_case : Dict = mapped_key.replace("*" , __lowerCamelCase ) if "weight_g" in name: snake_case : int = "weight_g" elif "weight_v" in name: snake_case : Dict = "weight_v" elif "bias" in name and "relative_attention_bias" not in name: snake_case : Dict = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case : List[Any] = "weight" else: snake_case : List[str] = None set_recursively(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) continue if not is_used: unused_weights.append(__lowerCamelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def UpperCamelCase ( __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : Any ): snake_case : Dict = full_name.split("conv_layers." )[-1] snake_case : List[Any] = name.split("." ) snake_case : Tuple = int(items[0] ) snake_case : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) snake_case : Optional[Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) snake_case : List[str] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) snake_case : str = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) snake_case : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__lowerCamelCase ) @torch.no_grad() def UpperCamelCase ( __lowerCamelCase : Dict , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any]=None ): # load the pre-trained checkpoints snake_case : Optional[int] = torch.load(__lowerCamelCase ) snake_case : List[Any] = WavLMConfigOrig(checkpoint["cfg"] ) snake_case : Optional[int] = WavLMOrig(__lowerCamelCase ) model.load_state_dict(checkpoint["model"] ) model.eval() if config_path is not None: snake_case : str = WavLMConfig.from_pretrained(__lowerCamelCase ) else: snake_case : Optional[int] = WavLMConfig() snake_case : Any = WavLMModel(__lowerCamelCase ) recursively_load_weights(__lowerCamelCase , __lowerCamelCase ) hf_wavlm.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") __lowerCamelCase = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
59
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case_( a__ , a__ , a__ , unittest.TestCase ): __UpperCamelCase = StableDiffusionInpaintPipeline __UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __UpperCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __UpperCamelCase = frozenset([] ) def lowerCamelCase__ ( self : Optional[Any] ): torch.manual_seed(0 ) lowerCAmelCase : Dict = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=UpperCamelCase_ , ) lowerCAmelCase : List[Any] = PNDMScheduler(skip_prk_steps=UpperCamelCase_ ) torch.manual_seed(0 ) lowerCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) lowerCAmelCase : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) lowerCAmelCase : Any = CLIPTextModel(UpperCamelCase_ ) lowerCAmelCase : str = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase : int = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Dict=0 ): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched lowerCAmelCase : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(UpperCamelCase_ ) ).to(UpperCamelCase_ ) lowerCAmelCase : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase : Union[str, Any] = Image.fromarray(np.uinta(UpperCamelCase_ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) lowerCAmelCase : Any = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((6_4, 6_4) ) if str(UpperCamelCase_ ).startswith('''mps''' ): lowerCAmelCase : Optional[Any] = torch.manual_seed(UpperCamelCase_ ) else: lowerCAmelCase : Dict = torch.Generator(device=UpperCamelCase_ ).manual_seed(UpperCamelCase_ ) lowerCAmelCase : Optional[int] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def lowerCamelCase__ ( self : Dict ): lowerCAmelCase : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase : Dict = self.get_dummy_components() lowerCAmelCase : Any = StableDiffusionInpaintPipeline(**UpperCamelCase_ ) lowerCAmelCase : Optional[Any] = sd_pipe.to(UpperCamelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase_ ) lowerCAmelCase : Optional[int] = self.get_dummy_inputs(UpperCamelCase_ ) lowerCAmelCase : Tuple = sd_pipe(**UpperCamelCase_ ).images lowerCAmelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowerCAmelCase : Optional[Any] = np.array([0.4_727, 0.5_735, 0.3_941, 0.5_446, 0.5_926, 0.4_394, 0.5_062, 0.4_654, 0.4_476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self : str ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class snake_case_( unittest.TestCase ): def lowerCamelCase__ ( self : List[Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self : str ): lowerCAmelCase : Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) lowerCAmelCase : Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) lowerCAmelCase : List[str] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) lowerCAmelCase : Union[str, Any] = '''stabilityai/stable-diffusion-2-inpainting''' lowerCAmelCase : Tuple = StableDiffusionInpaintPipeline.from_pretrained(UpperCamelCase_ , safety_checker=UpperCamelCase_ ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) pipe.enable_attention_slicing() lowerCAmelCase : int = '''Face of a yellow cat, high resolution, sitting on a park bench''' lowerCAmelCase : List[str] = torch.manual_seed(0 ) lowerCAmelCase : int = pipe( prompt=UpperCamelCase_ , image=UpperCamelCase_ , mask_image=UpperCamelCase_ , generator=UpperCamelCase_ , output_type='''np''' , ) lowerCAmelCase : Optional[int] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def lowerCamelCase__ ( self : Tuple ): lowerCAmelCase : Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) lowerCAmelCase : str = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) lowerCAmelCase : int = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) lowerCAmelCase : int = '''stabilityai/stable-diffusion-2-inpainting''' lowerCAmelCase : Any = StableDiffusionInpaintPipeline.from_pretrained( UpperCamelCase_ , torch_dtype=torch.floataa , safety_checker=UpperCamelCase_ , ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) pipe.enable_attention_slicing() lowerCAmelCase : str = '''Face of a yellow cat, high resolution, sitting on a park bench''' lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) lowerCAmelCase : Tuple = pipe( prompt=UpperCamelCase_ , image=UpperCamelCase_ , mask_image=UpperCamelCase_ , generator=UpperCamelCase_ , output_type='''np''' , ) lowerCAmelCase : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def lowerCamelCase__ ( self : Any ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCAmelCase : Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) lowerCAmelCase : Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) lowerCAmelCase : Union[str, Any] = '''stabilityai/stable-diffusion-2-inpainting''' lowerCAmelCase : List[str] = PNDMScheduler.from_pretrained(UpperCamelCase_ , subfolder='''scheduler''' ) lowerCAmelCase : Optional[int] = StableDiffusionInpaintPipeline.from_pretrained( UpperCamelCase_ , safety_checker=UpperCamelCase_ , scheduler=UpperCamelCase_ , torch_dtype=torch.floataa , ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCAmelCase : int = '''Face of a yellow cat, high resolution, sitting on a park bench''' lowerCAmelCase : Tuple = torch.manual_seed(0 ) lowerCAmelCase : Dict = pipe( prompt=UpperCamelCase_ , image=UpperCamelCase_ , mask_image=UpperCamelCase_ , generator=UpperCamelCase_ , num_inference_steps=2 , output_type='''np''' , ) lowerCAmelCase : List[str] = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
60
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
0