code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations __A : Union[str, Any] = [True] * 1_0_0_0_0_0_1 __A : Union[str, Any] = 2 while i * i <= 1_0_0_0_0_0_0: if seive[i]: for j in range(i * i, 1_0_0_0_0_0_1, i): __A : str = False i += 1 def __a ( A__ : int ): return seive[n] def __a ( A__ : int ): return any(digit in "02468" for digit in str(A__ ) ) def __a ( A__ : int = 1000000 ): SCREAMING_SNAKE_CASE = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(A__ ) and not contains_an_even_digit(A__ ): SCREAMING_SNAKE_CASE = str(A__ ) SCREAMING_SNAKE_CASE = [int(str_num[j:] + str_num[:j] ) for j in range(len(A__ ) )] if all(is_prime(A__ ) for i in list_nums ): result.append(A__ ) return result def __a ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f'{len(find_circular_primes()) = }')
16
"""simple docstring""" from __future__ import annotations def lowercase ( UpperCamelCase : list[float] ): """simple docstring""" if len(UpperCamelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) A__ : Union[str, Any] =nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
656
0
def lowerCamelCase( a__ ,a__): _SCREAMING_SNAKE_CASE =len(a__) print('''The following activities are selected:''') # The first activity is always selected _SCREAMING_SNAKE_CASE =0 print(a__ ,end=''',''') # Consider rest of the activities for j in range(a__): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(a__ ,end=''',''') _SCREAMING_SNAKE_CASE =j if __name__ == "__main__": import doctest doctest.testmod() snake_case_ : Dict = [1, 3, 0, 5, 8, 5] snake_case_ : List[Any] = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
691
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : Optional[Any] = { "configuration_mega": ["MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegaConfig", "MegaOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ "MEGA_PRETRAINED_MODEL_ARCHIVE_LIST", "MegaForCausalLM", "MegaForMaskedLM", "MegaForMultipleChoice", "MegaForQuestionAnswering", "MegaForSequenceClassification", "MegaForTokenClassification", "MegaModel", "MegaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
from abc import ABC, abstractmethod from argparse import ArgumentParser class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): '''simple docstring''' @staticmethod @abstractmethod def A ( lowercase : ArgumentParser ): '''simple docstring''' raise NotImplementedError() @abstractmethod def A ( self : Any ): '''simple docstring''' raise NotImplementedError()
686
"""simple docstring""" def lowercase ( UpperCamelCase : int ): """simple docstring""" if num <= 0: raise ValueError("Input must be a positive integer" ) A__ : Union[str, Any] =[True] * (num + 1) A__ : Union[str, Any] =2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCamelCase ): A__ : str =False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
656
0
"""simple docstring""" from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets A : List[Any] = "\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n" A : int = "\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n" A : List[Any] = "\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n \"accuracy\": Accuracy\n \"f1\": F1 score\n \"pearson\": Pearson Correlation\n \"spearmanr\": Spearman Correlation\n \"matthews_correlation\": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'stsb')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)})\n {'pearson': 1.0, 'spearmanr': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'cola')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' return float((preds == labels).mean() ) def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = simple_accuracy(_UpperCamelCase , _UpperCamelCase ) __lowerCAmelCase = float(fa_score(y_true=_UpperCamelCase , y_pred=_UpperCamelCase ) ) return { "accuracy": acc, "f1": fa, } def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = float(pearsonr(_UpperCamelCase , _UpperCamelCase )[0] ) __lowerCAmelCase = float(spearmanr(_UpperCamelCase , _UpperCamelCase )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): '''simple docstring''' def snake_case ( self ): if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), "references": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def snake_case ( self , __a , __a ): if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(UpperCamelCase__ , UpperCamelCase__ )} elif self.config_name == "stsb": return pearson_and_spearman(UpperCamelCase__ , UpperCamelCase__ ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(UpperCamelCase__ , UpperCamelCase__ ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(UpperCamelCase__ , UpperCamelCase__ )} else: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" )
636
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : List[Any] ): A__ : Tuple =torch.nn.Linear(10 , 10 ) A__ : List[str] =torch.optim.SGD(model.parameters() , 0.1 ) A__ : Union[str, Any] =Accelerator() A__ : str =accelerator.prepare(UpperCamelCase__ ) try: pickle.loads(pickle.dumps(UpperCamelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' @register_to_config def __init__( self : Dict ,A_ : int ,A_ : int ,A_ : int ,A_ : float ,A_ : int ,A_ : int ,A_ : int ,A_ : int ,A_ : str ,A_ : bool = False ,) -> List[str]: super().__init__() A = nn.Embedding(UpperCamelCase__ ,UpperCamelCase__ ) A = nn.Embedding(UpperCamelCase__ ,UpperCamelCase__ ) A = False A = nn.Dropout(p=UpperCamelCase__ ) A = TaConfig( vocab_size=UpperCamelCase__ ,d_model=UpperCamelCase__ ,num_heads=UpperCamelCase__ ,d_kv=UpperCamelCase__ ,d_ff=UpperCamelCase__ ,dropout_rate=UpperCamelCase__ ,feed_forward_proj=UpperCamelCase__ ,is_decoder=UpperCamelCase__ ,is_encoder_decoder=UpperCamelCase__ ,) A = nn.ModuleList() for lyr_num in range(UpperCamelCase__ ): A = TaBlock(UpperCamelCase__ ) self.encoders.append(UpperCamelCase__ ) A = TaLayerNorm(UpperCamelCase__ ) A = nn.Dropout(p=UpperCamelCase__ ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ,A_ : List[str] ,A_ : str ) -> List[Any]: A = self.token_embedder(UpperCamelCase__ ) A = encoder_input_tokens.shape[1] A = torch.arange(UpperCamelCase__ ,device=encoder_input_tokens.device ) x += self.position_encoding(UpperCamelCase__ ) A = self.dropout_pre(UpperCamelCase__ ) # inverted the attention mask A = encoder_input_tokens.size() A = self.get_extended_attention_mask(UpperCamelCase__ ,UpperCamelCase__ ) for lyr in self.encoders: A = lyr(UpperCamelCase__ ,UpperCamelCase__ )[0] A = self.layer_norm(UpperCamelCase__ ) return self.dropout_post(UpperCamelCase__ ), encoder_inputs_mask
91
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __A : Optional[int] = None __A : Union[str, Any] = logging.get_logger(__name__) __A : List[Any] = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __A : str = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } __A : List[str] = { "google/bigbird-roberta-base": 4_096, "google/bigbird-roberta-large": 4_096, "google/bigbird-base-trivia-itc": 4_096, } __A : Tuple = "▁" class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Dict = VOCAB_FILES_NAMES __magic_name__ : Any = PRETRAINED_VOCAB_FILES_MAP __magic_name__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ : List[Any] = BigBirdTokenizer __magic_name__ : Any = ["""input_ids""", """attention_mask"""] __magic_name__ : List[int] = [] def __init__( self : str , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : int="</s>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[Any]="[SEP]" , UpperCamelCase__ : List[Any]="[MASK]" , UpperCamelCase__ : str="[CLS]" , **UpperCamelCase__ : List[Any] , ): A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else bos_token A__ : Optional[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else eos_token A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else unk_token A__ : int =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else pad_token A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else cls_token A__ : List[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) A__ : List[Any] =vocab_file A__ : Optional[int] =False if not self.vocab_file else True def _UpperCAmelCase ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1] def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : Dict =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : List[str] =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
656
0
"""simple docstring""" import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A_ = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right A_ = 50003 A_ = 50002 @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Any = PLBartTokenizer __lowerCamelCase : str = None __lowerCamelCase : Union[str, Any] = False def UpperCamelCase__ ( self: Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase_ =PLBartTokenizer(UpperCamelCase__ , language_codes="base" , keep_accents=UpperCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self: Union[str, Any] ): UpperCamelCase_ =PLBartTokenizer(UpperCamelCase__ , language_codes="base" , keep_accents=UpperCamelCase__ ) UpperCamelCase_ =tokenizer.tokenize("This is a test" ) self.assertListEqual(UpperCamelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCamelCase_ =tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) UpperCamelCase_ =tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) UpperCamelCase_ =tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) UpperCamelCase_ =tokenizer.vocab_size UpperCamelCase_ =[tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) for x in range(end - 4 , UpperCamelCase__ )] self.assertListEqual(UpperCamelCase__ , ["__java__", "__python__", "__en_XX__", "<mask>"] ) UpperCamelCase_ ="java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" UpperCamelCase_ =tokenizer(UpperCamelCase__ ).input_ids self.assertEqual( tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) , UpperCamelCase__ , ) def UpperCamelCase__ ( self: int ): UpperCamelCase_ =PLBartTokenizer(UpperCamelCase__ , language_codes="multi" , keep_accents=UpperCamelCase__ ) UpperCamelCase_ =tokenizer.tokenize("This is a test" ) self.assertListEqual(UpperCamelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCamelCase_ =tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) UpperCamelCase_ =tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) UpperCamelCase_ =tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) UpperCamelCase_ =tokenizer.vocab_size UpperCamelCase_ =[tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) for x in range(end - 7 , UpperCamelCase__ )] self.assertListEqual( UpperCamelCase__ , ["__java__", "__python__", "__en_XX__", "__javascript__", "__php__", "__ruby__", "__go__"] ) UpperCamelCase_ ="java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" UpperCamelCase_ =tokenizer(UpperCamelCase__ ).input_ids self.assertEqual( tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) , UpperCamelCase__ , ) @require_torch @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Optional[Any] = """uclanlp/plbart-python-en_XX""" __lowerCamelCase : List[str] = [ """def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])""", """def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])""", ] __lowerCamelCase : Tuple = [ """Returns the maximum value of a b c.""", """Sums the values of a b c.""", ] __lowerCamelCase : Dict = [ 1_34, 54_52, 3_34_60, 3_34_41, 3_34_63, 3_34_65, 3_34_63, 3_34_49, 9_88, 20, 3_34_56, 19, 3_34_56, 7_71, 39, 42_58, 8_89, 33_18, 3_34_41, 3_34_63, 3_34_65, 3_34_63, 3_34_49, 24_71, 2, PYTHON_CODE, ] @classmethod def UpperCamelCase__ ( cls: str ): UpperCamelCase_ =PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes="base" , src_lang="python" , tgt_lang="en_XX" ) UpperCamelCase_ =1 return cls def UpperCamelCase__ ( self: str ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__java__"] , 5_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__python__"] , 5_0002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__en_XX__"] , 5_0003 ) def UpperCamelCase__ ( self: Dict ): UpperCamelCase_ =self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase__ ) def UpperCamelCase__ ( self: Any ): self.assertIn(UpperCamelCase__ , self.tokenizer.all_special_ids ) UpperCamelCase_ =[EN_CODE, 9037, 3_3442, 57, 752, 153, 14, 56, 18, 9, 2] UpperCamelCase_ =self.tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) UpperCamelCase_ =self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase__ ) def UpperCamelCase__ ( self: Any ): UpperCamelCase_ =["def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])" * 20] self.assertIsInstance(src_text[0] , UpperCamelCase__ ) UpperCamelCase_ =10 UpperCamelCase_ =self.tokenizer(UpperCamelCase__ , max_length=UpperCamelCase__ , truncation=UpperCamelCase__ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , UpperCamelCase__ ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) def UpperCamelCase__ ( self: int ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "__java__"] ) , [5_0004, 5_0001] ) def UpperCamelCase__ ( self: Dict ): UpperCamelCase_ =tempfile.mkdtemp() UpperCamelCase_ =self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase__ ) UpperCamelCase_ =PLBartTokenizer.from_pretrained(UpperCamelCase__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase__ ) @require_torch def UpperCamelCase__ ( self: str ): UpperCamelCase_ =self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase__ , return_tensors="pt" ) UpperCamelCase_ =shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , UpperCamelCase__ ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def UpperCamelCase__ ( self: Dict ): UpperCamelCase_ =self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) UpperCamelCase_ =shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual((2, 26) , batch.input_ids.shape ) self.assertEqual((2, 26) , batch.attention_mask.shape ) UpperCamelCase_ =batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase__ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def UpperCamelCase__ ( self: Optional[int] ): UpperCamelCase_ =self.tokenizer(self.src_text , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=3 , return_tensors="pt" ) UpperCamelCase_ =self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=10 , return_tensors="pt" ) UpperCamelCase_ =targets["input_ids"] UpperCamelCase_ =shift_tokens_right(UpperCamelCase__ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def UpperCamelCase__ ( self: Any ): UpperCamelCase_ =self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="java" ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , { # A, test, EOS, en_XX "input_ids": [[150, 242, 2, 5_0003]], "attention_mask": [[1, 1, 1, 1]], # java "forced_bos_token_id": 5_0001, } , )
391
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Optional[int] = logging.get_logger(__name__) __A : Optional[int] = {"vocab_file": "spiece.model"} __A : List[Any] = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : Dict="<s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : Optional[int]="<sep>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[int]="<cls>" , UpperCamelCase__ : List[str]="<mask>" , UpperCamelCase__ : Optional[Any]=["<eop>", "<eod>"] , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : Dict , ): A__ : List[str] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token A__ : Tuple ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) A__ : Dict =3 A__ : int =do_lower_case A__ : str =remove_space A__ : Optional[Any] =keep_accents A__ : int =vocab_file A__ : Dict =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) A__ : Union[str, Any] =jieba A__ : List[str] =str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _UpperCAmelCase ( self : Union[str, Any] ): return len(self.sp_model ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Any ={self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[str] ): A__ : Union[str, Any] =self.__dict__.copy() A__ : Tuple =None return state def __setstate__( self : Tuple , UpperCamelCase__ : int ): A__ : Union[str, Any] =d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A__ : Optional[int] ={} A__ : Union[str, Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Dict ): if self.remove_space: A__ : Optional[int] =" ".join(inputs.strip().split() ) else: A__ : Optional[Any] =inputs A__ : Any =outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: A__ : Optional[Any] =unicodedata.normalize("NFKD" , UpperCamelCase__ ) A__ : Tuple ="".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: A__ : str =outputs.lower() return outputs def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : str ): A__ : Optional[int] =self.preprocess_text(UpperCamelCase__ ) A__ : Dict =self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) A__ : List[str] =[] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): A__ : str =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : Union[str, Any] =cur_pieces[1:] else: A__ : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def _UpperCAmelCase ( self : int , UpperCamelCase__ : str ): return self.sp_model.PieceToId(UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[Any] ): return self.sp_model.IdToPiece(UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : str ): A__ : Optional[int] ="".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def _UpperCAmelCase ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : Optional[Any] =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : Tuple =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: A__ : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def _UpperCAmelCase ( self : str , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : int ): A__ : List[Any] =super()._decode(*UpperCamelCase__ , **UpperCamelCase__ ) A__ : Union[str, Any] =text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
656
0
'''simple docstring''' import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def a_ ( UpperCamelCase_ ): if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : str , lowerCAmelCase : nn.Module , lowerCAmelCase : int ): super().__init__() A_ = module A_ = nn.Sequential( nn.Linear(module.in_features , UpperCamelCase__ , bias=UpperCamelCase__ ) , nn.Linear(UpperCamelCase__ , module.out_features , bias=UpperCamelCase__ ) , ) A_ = (2.0 / (5 * min(module.in_features , module.out_features ))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=UpperCamelCase__ ) nn.init.zeros_(self.adapter[1].weight ) self.adapter.to(module.weight.device ) def _UpperCAmelCase ( self : List[str] , lowerCAmelCase : Optional[Any] , *lowerCAmelCase : Union[str, Any] , **lowerCAmelCase : Union[str, Any] ): return self.module(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) + self.adapter(UpperCamelCase__ ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" _UpperCAmelCase : str ="""bigscience/bloom-1b7""" # Constant values _UpperCAmelCase : List[str] =2.1_0_9_6_5_9_5_5_2_6_9_2_5_7_4 _UpperCAmelCase : Tuple ="""Hello my name is""" _UpperCAmelCase : int =set() EXPECTED_OUTPUTS.add("Hello my name is John and I am a professional photographer. I" ) EXPECTED_OUTPUTS.add("Hello my name is John.\nI am a friend of your father.\n" ) EXPECTED_OUTPUTS.add("Hello my name is John Doe, I am a student at the University" ) _UpperCAmelCase : int =10 def _UpperCAmelCase ( self : int ): # Models and tokenizer A_ = AutoTokenizer.from_pretrained(self.model_name ) class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : Any ): super().setUp() # Models and tokenizer A_ = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map="auto" ) A_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) def _UpperCAmelCase ( self : Union[str, Any] ): del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self : int ): A_ = self.model_abit.config self.assertTrue(hasattr(UpperCamelCase__ , "quantization_config" ) ) A_ = config.to_dict() A_ = config.to_diff_dict() A_ = config.to_json_string() def _UpperCAmelCase ( self : str ): from bitsandbytes.nn import Paramsabit A_ = self.model_fpaa.get_memory_footprint() A_ = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE ) A_ = get_some_linear_layer(self.model_abit ) self.assertTrue(linear.weight.__class__ == Paramsabit ) def _UpperCAmelCase ( self : List[Any] ): from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(UpperCamelCase__ , torch.nn.Linear ): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta ) def _UpperCAmelCase ( self : int ): A_ = self.tokenizer(self.input_text , return_tensors="pt" ) A_ = self.model_abit.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=UpperCamelCase__ ) , self.EXPECTED_OUTPUTS ) def _UpperCAmelCase ( self : List[str] ): A_ = BitsAndBytesConfig() A_ = True A_ = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=UpperCamelCase__ , device_map="auto" ) A_ = self.tokenizer(self.input_text , return_tensors="pt" ) A_ = model_abit_from_config.generate( input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=UpperCamelCase__ ) , self.EXPECTED_OUTPUTS ) def _UpperCAmelCase ( self : Optional[int] ): with self.assertRaises(UpperCamelCase__ ), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(UpperCamelCase__ ) def _UpperCAmelCase ( self : List[Any] ): A_ = BitsAndBytesConfig() with self.assertRaises(UpperCamelCase__ ): A_ = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=UpperCamelCase__ , load_in_abit=UpperCamelCase__ , device_map="auto" , bnb_abit_quant_type="nf4" , ) def _UpperCAmelCase ( self : str ): with self.assertRaises(UpperCamelCase__ ): # Tries with `str` self.model_abit.to("cpu" ) with self.assertRaises(UpperCamelCase__ ): # Tries with a `dtype`` self.model_abit.to(torch.floataa ) with self.assertRaises(UpperCamelCase__ ): # Tries with a `device` self.model_abit.to(torch.device("cuda:0" ) ) with self.assertRaises(UpperCamelCase__ ): # Tries with a `device` self.model_abit.float() with self.assertRaises(UpperCamelCase__ ): # Tries with a `device` self.model_abit.half() # Test if we did not break anything A_ = self.tokenizer(self.input_text , return_tensors="pt" ) A_ = self.model_fpaa.to(torch.floataa ) A_ = self.model_fpaa.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) # Check this does not throw an error A_ = self.model_fpaa.to("cpu" ) # Check this does not throw an error A_ = self.model_fpaa.half() # Check this does not throw an error A_ = self.model_fpaa.float() def _UpperCAmelCase ( self : Optional[Any] ): A_ = AutoModelForSeqaSeqLM.from_pretrained("t5-small" , load_in_abit=UpperCamelCase__ , device_map="auto" ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @classmethod def _UpperCAmelCase ( cls : Dict ): A_ = "t5-small" A_ = "google/flan-t5-small" # flan-t5 uses dense-act instead of dense-relu-dense A_ = AutoTokenizer.from_pretrained(cls.model_name ) A_ = "Translate in German: Hello, my dog is cute" def _UpperCAmelCase ( self : Dict ): gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self : Tuple ): from transformers import TaForConditionalGeneration A_ = TaForConditionalGeneration._keep_in_fpaa_modules A_ = None # test with `t5-small` A_ = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) A_ = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) A_ = model.generate(**UpperCamelCase__ ) # test with `flan-t5-small` A_ = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) A_ = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) A_ = model.generate(**UpperCamelCase__ ) A_ = modules def _UpperCAmelCase ( self : Tuple ): import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` A_ = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit ) ) A_ = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) A_ = model.generate(**UpperCamelCase__ ) # test with `flan-t5-small` A_ = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) A_ = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) A_ = model.generate(**UpperCamelCase__ ) class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : Optional[int] ): super().setUp() # model_name A_ = "bigscience/bloom-560m" A_ = "t5-small" # Different types of model A_ = AutoModel.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) # Sequence classification model A_ = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) # CausalLM model A_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) # Seq2seq model A_ = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=UpperCamelCase__ , device_map="auto" ) def _UpperCAmelCase ( self : int ): del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self : List[Any] ): from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit ) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter ) class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : str ): super().setUp() def _UpperCAmelCase ( self : Optional[int] ): del self.pipe gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self : str ): A_ = pipeline( "text-generation" , model=self.model_name , model_kwargs={"device_map": "auto", "load_in_4bit": True, "torch_dtype": torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass A_ = self.pipe(self.input_text ) self.assertIn(pipeline_output[0]["generated_text"] , self.EXPECTED_OUTPUTS ) @require_torch_multi_gpu class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : List[Any] ): super().setUp() def _UpperCAmelCase ( self : List[Any] ): A_ = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=UpperCamelCase__ , device_map="balanced" ) # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values() ) , {0, 1} ) # Check that inference pass works on the model A_ = self.tokenizer(self.input_text , return_tensors="pt" ) # Second real batch A_ = model_parallel.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=UpperCamelCase__ ) , self.EXPECTED_OUTPUTS ) class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : Tuple ): A_ = "facebook/opt-350m" super().setUp() def _UpperCAmelCase ( self : List[Any] ): if version.parse(importlib.metadata.version("bitsandbytes" ) ) < version.parse("0.37.0" ): return # Step 1: freeze all parameters A_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase__ ) self.assertEqual(set(model.hf_device_map.values() ) , {torch.cuda.current_device()} ) for param in model.parameters(): A_ = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability A_ = param.data.to(torch.floataa ) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(UpperCamelCase__ ) ): A_ = LoRALayer(module.q_proj , rank=16 ) A_ = LoRALayer(module.k_proj , rank=16 ) A_ = LoRALayer(module.v_proj , rank=16 ) # Step 3: dummy batch A_ = self.tokenizer("Test batch " , return_tensors="pt" ).to(0 ) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): A_ = model.forward(**UpperCamelCase__ ) out.logits.norm().backward() for module in model.modules(): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): self.assertTrue(module.adapter[1].weight.grad is not None ) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0 ) elif isinstance(UpperCamelCase__ , nn.Embedding ): self.assertTrue(module.weight.grad is None ) class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" _UpperCAmelCase : List[str] ="""gpt2-xl""" _UpperCAmelCase : List[Any] =3.3_1_9_1_8_5_4_8_5_4_1_5_2_1_8_7
452
"""simple docstring""" def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations(UpperCamelCase : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations_with_dp_array( UpperCamelCase : int , UpperCamelCase : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] A__ : str =sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase ) for item in array ) A__ : List[str] =answer return answer A__ : List[Any] =[-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase , UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" A__ : str =[0] * (target + 1) A__ : Optional[Any] =1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[Any] = 3 __A : Optional[Any] = 5 __A : int = [1, 2, 5] print(combination_sum_iv(n, array, target))
656
0
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def a_ ( lowercase__ :str="ro", lowercase__ :Tuple="en", lowercase__ :Tuple="wmt16", lowercase__ :Any=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("""run pip install datasets""" ) __lowerCamelCase = f'{src_lang}-{tgt_lang}' print(f'Converting {dataset}-{pair}' ) __lowerCamelCase = datasets.load_dataset(lowercase__, lowercase__ ) if save_dir is None: __lowerCamelCase = f'{dataset}-{pair}' __lowerCamelCase = Path(lowercase__ ) save_dir.mkdir(exist_ok=lowercase__ ) for split in ds.keys(): print(f'Splitting {split} with {ds[split].num_rows} records' ) # to save to val.source, val.target like summary datasets __lowerCamelCase = "val" if split == "validation" else split __lowerCamelCase = save_dir.joinpath(f'{fn}.source' ) __lowerCamelCase = save_dir.joinpath(f'{fn}.target' ) __lowerCamelCase = src_path.open("""w+""" ) __lowerCamelCase = tgt_path.open("""w+""" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): __lowerCamelCase = x["translation"] src_fp.write(ex[src_lang] + """\n""" ) tgt_fp.write(ex[tgt_lang] + """\n""" ) print(f'Saved {dataset} dataset to {save_dir}' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
281
"""simple docstring""" import math import tensorflow as tf from packaging import version def lowercase ( UpperCamelCase : Optional[Any] ): """simple docstring""" A__ : List[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : Optional[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : Tuple =tf.cast(math.pi , x.dtype ) A__ : Dict =tf.cast(0.04_47_15 , x.dtype ) A__ : Optional[int] =0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(UpperCamelCase , 3 )) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) return x * tf.tanh(tf.math.softplus(UpperCamelCase ) ) def lowercase ( UpperCamelCase : List[str] ): """simple docstring""" A__ : Union[str, Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =tf.cast(0.04_47_15 , x.dtype ) A__ : List[Any] =tf.cast(0.79_78_84_56_08 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) A__ : str =tf.cast(1.7_02 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" return tf.clip_by_value(_gelu(UpperCamelCase ) , -10 , 10 ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any=-1 ): """simple docstring""" A__ , A__ : Optional[Any] =tf.split(UpperCamelCase , 2 , axis=UpperCamelCase ) return a * tf.math.sigmoid(UpperCamelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def lowercase ( UpperCamelCase : int ): """simple docstring""" return tf.keras.activations.gelu(UpperCamelCase , approximate=UpperCamelCase ) __A : Optional[Any] = tf.keras.activations.gelu __A : Optional[Any] = approximate_gelu_wrap else: __A : Any = _gelu __A : Union[str, Any] = _gelu_new __A : List[str] = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F'''function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}''' )
656
0
import math import tensorflow as tf from packaging import version def a__ (__lowercase :Optional[Any] ) -> List[Any]: _A : List[Any] = tf.convert_to_tensor(__lowercase ) _A : List[Any] = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def a__ (__lowercase :Optional[int] ) -> Dict: _A : Optional[Any] = tf.convert_to_tensor(__lowercase ) _A : Tuple = tf.cast(math.pi , x.dtype ) _A : Dict = tf.cast(0.04_4715 , x.dtype ) _A : Optional[int] = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(__lowercase , 3 )) )) return x * cdf def a__ (__lowercase :Optional[int] ) -> Optional[int]: _A : List[str] = tf.convert_to_tensor(__lowercase ) return x * tf.tanh(tf.math.softplus(__lowercase ) ) def a__ (__lowercase :List[str] ) -> Union[str, Any]: _A : Union[str, Any] = tf.convert_to_tensor(__lowercase ) _A : List[Any] = tf.cast(0.04_4715 , x.dtype ) _A : List[Any] = tf.cast(0.79_7884_5608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def a__ (__lowercase :List[Any] ) -> Optional[Any]: _A : List[str] = tf.convert_to_tensor(__lowercase ) _A : str = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def a__ (__lowercase :Tuple ) -> List[str]: return tf.clip_by_value(_gelu(__lowercase ) , -10 , 10 ) def a__ (__lowercase :str , __lowercase :Any=-1 ) -> Optional[int]: _A : Optional[Any] = tf.split(__lowercase , 2 , axis=__lowercase ) return a * tf.math.sigmoid(__lowercase ) if version.parse(tf.version.VERSION) >= version.parse('2.4'): def a__ (__lowercase :int ) -> Union[str, Any]: return tf.keras.activations.gelu(__lowercase , approximate=__lowercase ) _UpperCamelCase : Optional[Any] =tf.keras.activations.gelu _UpperCamelCase : Optional[Any] =approximate_gelu_wrap else: _UpperCamelCase : Any =_gelu _UpperCamelCase : Union[str, Any] =_gelu_new _UpperCamelCase : List[str] ={ "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def a__ (__lowercase :List[Any] ) -> Optional[int]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(f"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
206
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict ): A__ : Optional[Any] =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_attention_heads" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_encoder_blocks" ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any]=13 , UpperCamelCase__ : Tuple=64 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Union[str, Any]=4 , UpperCamelCase__ : Dict=[2, 2, 2, 2] , UpperCamelCase__ : Union[str, Any]=[8, 4, 2, 1] , UpperCamelCase__ : Tuple=[16, 32, 64, 128] , UpperCamelCase__ : Optional[int]=[1, 4, 8, 16] , UpperCamelCase__ : Any=[1, 2, 4, 8] , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Dict="gelu" , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[Any]=None , ): A__ : Tuple =parent A__ : List[Any] =batch_size A__ : List[Any] =image_size A__ : Union[str, Any] =num_channels A__ : Optional[int] =num_encoder_blocks A__ : Any =sr_ratios A__ : Any =depths A__ : List[Any] =hidden_sizes A__ : List[Any] =downsampling_rates A__ : List[str] =num_attention_heads A__ : int =is_training A__ : List[Any] =use_labels A__ : Any =hidden_act A__ : Dict =hidden_dropout_prob A__ : int =attention_probs_dropout_prob A__ : List[Any] =initializer_range A__ : Tuple =num_labels A__ : List[Any] =scope def _UpperCAmelCase ( self : Optional[int] ): A__ : List[str] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Any =None if self.use_labels: A__ : Tuple =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A__ : List[Any] =self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self : Tuple ): return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int ): A__ : Any =SegformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Dict =model(UpperCamelCase__ ) A__ : Optional[int] =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[Any] ): A__ : str =self.num_labels A__ : Optional[Any] =SegformerForSemanticSegmentation(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Optional[Any] =model(UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) A__ : List[Any] =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str ): A__ : Tuple =1 A__ : Tuple =SegformerForSemanticSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : List[str] =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(UpperCamelCase__ ) A__ : Dict =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : str ): A__ : Union[str, Any] =self.prepare_config_and_inputs() A__ , A__ , A__ : Tuple =config_and_inputs A__ : Tuple ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Dict = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __magic_name__ : Optional[int] = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ : Dict = True __magic_name__ : List[str] = False __magic_name__ : Optional[Any] = False __magic_name__ : str = False def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =SegformerModelTester(self ) A__ : Tuple =SegformerConfigTester(self , config_class=UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): self.config_tester.run_common_tests() def _UpperCAmelCase ( self : Dict ): A__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*UpperCamelCase__ ) @unittest.skip("SegFormer does not use inputs_embeds" ) def _UpperCAmelCase ( self : Dict ): pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods" ) def _UpperCAmelCase ( self : Tuple ): pass def _UpperCAmelCase ( self : List[str] ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : int =model_class(UpperCamelCase__ ) A__ : Optional[int] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Optional[int] =[*signature.parameters.keys()] A__ : List[str] =["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() A__ : Union[str, Any] =True for model_class in self.all_model_classes: A__ : Optional[Any] =True A__ : Union[str, Any] =False A__ : str =True A__ : Optional[int] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : str =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Any =outputs.attentions A__ : List[str] =sum(self.model_tester.depths ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ : Dict =True A__ : str =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Any =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Union[str, Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : List[Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) A__ : Tuple =(self.model_tester.image_size // 32) ** 2 A__ : Optional[Any] =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) A__ : int =len(UpperCamelCase__ ) # Check attention is always last and order is fine A__ : Optional[Any] =True A__ : Any =True A__ : Union[str, Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Optional[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(out_len + 1 , len(UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : Union[str, Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _UpperCAmelCase ( self : List[Any] ): def check_hidden_states_output(UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple ): A__ : Optional[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : List[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.hidden_states A__ : int =self.model_tester.num_encoder_blocks self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Optional[Any] =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : str =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[int] ): if not self.model_tester.is_training: return A__ , A__ : int =self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] =True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ): continue A__ : List[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() A__ : int =self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) A__ : Union[str, Any] =model(**UpperCamelCase__ ).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _UpperCAmelCase ( self : Tuple ): pass @slow def _UpperCAmelCase ( self : Tuple ): for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Tuple =SegformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowercase ( ): """simple docstring""" A__ : List[Any] =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): # only resize + normalize A__ : List[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : Union[str, Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : Union[str, Any] =prepare_img() A__ : Union[str, Any] =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : int =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : Dict =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : Optional[int] =torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : Union[str, Any] ): # only resize + normalize A__ : Dict =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : int =SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024" ).to(UpperCamelCase__ ) A__ : Tuple =prepare_img() A__ : str =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Optional[int] =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : List[str] =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : List[Any] =torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-1 ) ) @slow def _UpperCAmelCase ( self : int ): # only resize + normalize A__ : Optional[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : List[Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : str =prepare_img() A__ : Dict =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Any =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : Dict =model(UpperCamelCase__ ) A__ : Any =outputs.logits.detach().cpu() A__ : Union[str, Any] =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ , target_sizes=[(500, 300)] ) A__ : List[str] =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ ) A__ : int =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ ) A__ : Tuple =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ )
656
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { "facebook/convnextv2-tiny-1k-224": "https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json", } class __magic_name__ ( _UpperCamelCase , _UpperCamelCase ): _SCREAMING_SNAKE_CASE : List[Any] = """convnextv2""" def __init__( self : Any , snake_case_ : Union[str, Any]=3 , snake_case_ : List[Any]=4 , snake_case_ : Dict=4 , snake_case_ : List[Any]=None , snake_case_ : Dict=None , snake_case_ : str="gelu" , snake_case_ : Any=0.02 , snake_case_ : str=1e-12 , snake_case_ : Optional[Any]=0.0 , snake_case_ : int=224 , snake_case_ : Optional[int]=None , snake_case_ : Optional[Any]=None , **snake_case_ : str , ): super().__init__(**UpperCamelCase__ ) __snake_case = num_channels __snake_case = patch_size __snake_case = num_stages __snake_case = [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes __snake_case = [3, 3, 9, 3] if depths is None else depths __snake_case = hidden_act __snake_case = initializer_range __snake_case = layer_norm_eps __snake_case = drop_path_rate __snake_case = image_size __snake_case = ["stem"] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] __snake_case = get_aligned_output_features_output_indices( out_features=UpperCamelCase__ , out_indices=UpperCamelCase__ , stage_names=self.stage_names )
163
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any]=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[str]=99 , UpperCamelCase__ : Optional[Any]=32 , UpperCamelCase__ : Any=5 , UpperCamelCase__ : Dict=4 , UpperCamelCase__ : Union[str, Any]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Any=0.1 , UpperCamelCase__ : Optional[Any]=0.1 , UpperCamelCase__ : Optional[Any]=512 , UpperCamelCase__ : Union[str, Any]=16 , UpperCamelCase__ : List[str]=2 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : List[Any]=4 , ): A__ : str =parent A__ : List[str] =batch_size A__ : Any =seq_length A__ : List[str] =is_training A__ : List[Any] =use_attention_mask A__ : List[Any] =use_token_type_ids A__ : Dict =use_labels A__ : List[Any] =vocab_size A__ : Optional[int] =hidden_size A__ : Optional[Any] =num_hidden_layers A__ : str =num_attention_heads A__ : int =intermediate_size A__ : Tuple =hidden_act A__ : Tuple =hidden_dropout_prob A__ : Dict =attention_probs_dropout_prob A__ : Any =max_position_embeddings A__ : Any =type_vocab_size A__ : Union[str, Any] =type_sequence_label_size A__ : Optional[Any] =initializer_range A__ : int =num_choices def _UpperCAmelCase ( self : Tuple ): A__ : List[str] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[str] =None if self.use_attention_mask: A__ : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) A__ : str =None if self.use_token_type_ids: A__ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Any =RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCAmelCase ( self : Tuple ): A__ : Dict =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : str =config_and_inputs A__ : Optional[Any] ={"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _UpperCAmelCase ( self : int ): A__ : str =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Union[str, Any] =config_and_inputs A__ : Union[str, Any] =True A__ : List[Any] =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : Tuple =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __lowerCAmelCase ( _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Union[str, Any] = True __magic_name__ : Dict = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Optional[int] =FlaxRobertaPreLayerNormModelTester(self ) @slow def _UpperCAmelCase ( self : List[Any] ): for model_class_name in self.all_model_classes: A__ : Tuple =model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Union[str, Any] =model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): A__ : Any =FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Tuple =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : str =model(UpperCamelCase__ )[0] A__ : List[Any] =[1, 11, 50265] self.assertEqual(list(output.shape ) , UpperCamelCase__ ) # compare the actual values for a slice. A__ : Any =np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : List[Any] ): A__ : Union[str, Any] =FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : List[Any] =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : Dict =model(UpperCamelCase__ )[0] # compare the actual values for a slice. A__ : Optional[Any] =np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) )
656
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor SCREAMING_SNAKE_CASE__ : Optional[Any] =logging.get_logger(__name__) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self , *_lowercase , **_lowercase ) -> List[Any]: warnings.warn( '''The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PoolFormerImageProcessor instead.''' , UpperCamelCase__ , ) super().__init__(*UpperCamelCase__ , **UpperCamelCase__ )
434
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __A : List[Any] = logging.get_logger(__name__) __A : Any = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] __A : Optional[int] = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" A__ : Union[str, Any] =torch.load(UpperCamelCase , map_location="cpu" ) return sd def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : int=rename_keys_prefix ): """simple docstring""" A__ : List[str] =OrderedDict() A__ : str =torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue A__ : Optional[Any] =key for name_pair in rename_keys_prefix: A__ : int =new_key.replace(name_pair[0] , name_pair[1] ) A__ : Dict =d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately A__ : Optional[int] =new_d["cls.predictions.bias"] return new_d @torch.no_grad() def lowercase ( UpperCamelCase : Dict , UpperCamelCase : List[str] ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: A__ : Any ="pretraining" if "vcr" in checkpoint_path: A__ : Union[str, Any] ={"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: A__ : Optional[Any] ={"visual_embedding_dim": 2048} elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048} elif "nlvr" in checkpoint_path: A__ : List[str] ={"visual_embedding_dim": 1024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 512} A__ : List[str] ="multichoice" elif "vqa_advanced" in checkpoint_path: A__ : Any ={"visual_embedding_dim": 2048} A__ : str ="vqa_advanced" elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048, "num_labels": 3129} A__ : str ="vqa" elif "nlvr" in checkpoint_path: A__ : str ={ "visual_embedding_dim": 1024, "num_labels": 2, } A__ : Dict ="nlvr" A__ : Union[str, Any] =VisualBertConfig(**UpperCamelCase ) # Load State Dict A__ : int =load_state_dict(UpperCamelCase ) A__ : Tuple =get_new_dict(UpperCamelCase , UpperCamelCase ) if model_type == "pretraining": A__ : str =VisualBertForPreTraining(UpperCamelCase ) elif model_type == "vqa": A__ : Optional[int] =VisualBertForQuestionAnswering(UpperCamelCase ) elif model_type == "nlvr": A__ : Union[str, Any] =VisualBertForVisualReasoning(UpperCamelCase ) elif model_type == "multichoice": A__ : Union[str, Any] =VisualBertForMultipleChoice(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) # Save Checkpoints Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": __A : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("orig_checkpoint_path", type=str, help="A path to .th on local filesystem.") parser.add_argument("pytorch_dump_folder_path", type=str, help="Path to the output PyTorch model.") __A : str = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
656
0
from ....configuration_utils import PretrainedConfig from ....utils import logging __A : Any = logging.get_logger(__name__) # TODO: upload to AWS __A : int = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json" ), } class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' lowerCamelCase__ = """retribert""" def __init__( self : Any , __lowerCamelCase : Tuple=30522 , __lowerCamelCase : Any=768 , __lowerCamelCase : List[str]=8 , __lowerCamelCase : List[Any]=12 , __lowerCamelCase : Optional[Any]=3072 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : Any=512 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : str=0.02 , __lowerCamelCase : List[Any]=1e-12 , __lowerCamelCase : Any=True , __lowerCamelCase : Dict=128 , __lowerCamelCase : int=0 , **__lowerCamelCase : Optional[Any] , ): super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = type_vocab_size SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = layer_norm_eps SCREAMING_SNAKE_CASE = share_encoders SCREAMING_SNAKE_CASE = projection_dim
16
"""simple docstring""" __A : Union[str, Any] = {str(digit): digit**5 for digit in range(10)} def lowercase ( UpperCamelCase : int ): """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase ) ) def lowercase ( ): """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(UpperCamelCase ) ) if __name__ == "__main__": print(solution())
656
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case_ : Dict = logging.get_logger(__name__) snake_case_ : Dict = "▁" snake_case_ : Tuple = {"vocab_file": "sentencepiece.bpe.model"} snake_case_ : Union[str, Any] = { "vocab_file": { "facebook/xglm-564M": "https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model", } } snake_case_ : Optional[int] = { "facebook/xglm-564M": 20_48, } class A__ ( _UpperCamelCase ): UpperCAmelCase = VOCAB_FILES_NAMES UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase = ["""input_ids""", """attention_mask"""] def __init__( self : Dict , _a : List[str] , _a : Optional[Any]="<s>" , _a : str="</s>" , _a : Optional[int]="</s>" , _a : List[Any]="<s>" , _a : int="<unk>" , _a : Tuple="<pad>" , _a : Optional[Dict[str, Any]] = None , **_a : List[str] , ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE ={} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer _SCREAMING_SNAKE_CASE =7 _SCREAMING_SNAKE_CASE =[f"<madeupword{i}>" for i in range(self.num_madeup_words )] _SCREAMING_SNAKE_CASE =kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) _SCREAMING_SNAKE_CASE =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase__ ) ) _SCREAMING_SNAKE_CASE =vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _SCREAMING_SNAKE_CASE =1 # Mimic fairseq token-to-id alignment for the first 4 token _SCREAMING_SNAKE_CASE ={"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} _SCREAMING_SNAKE_CASE =len(self.sp_model ) _SCREAMING_SNAKE_CASE ={f"<madeupword{i}>": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE ={v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =self.__dict__.copy() _SCREAMING_SNAKE_CASE =None _SCREAMING_SNAKE_CASE =self.sp_model.serialized_model_proto() return state def __setstate__( self : List[Any] , _a : Optional[int] ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _SCREAMING_SNAKE_CASE ={} _SCREAMING_SNAKE_CASE =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __UpperCamelCase ( self : Optional[Any] , _a : List[int] , _a : Optional[List[int]] = None ) -> Any: """simple docstring""" if token_ids_a is None: return [self.sep_token_id] + token_ids_a _SCREAMING_SNAKE_CASE =[self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def __UpperCamelCase ( self : int , _a : List[int] , _a : Optional[List[int]] = None , _a : bool = False ) -> Tuple: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) return [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] + ([0] * len(UpperCamelCase__ )) def __UpperCamelCase ( self : Union[str, Any] , _a : List[int] , _a : Optional[List[int]] = None ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =[self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def __UpperCamelCase ( self : List[str] ) -> Dict: """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def __UpperCamelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE ={self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __UpperCamelCase ( self : Optional[int] , _a : str ) -> Tuple: """simple docstring""" return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def __UpperCamelCase ( self : Optional[Any] , _a : Any ) -> Optional[Any]: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _SCREAMING_SNAKE_CASE =self.sp_model.PieceToId(UpperCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __UpperCamelCase ( self : Optional[Any] , _a : str ) -> str: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __UpperCamelCase ( self : int , _a : str ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE ="".join(UpperCamelCase__ ).replace(UpperCamelCase__ , ''' ''' ).strip() return out_string def __UpperCamelCase ( self : int , _a : str , _a : Optional[str] = None ) -> List[str]: """simple docstring""" if not os.path.isdir(UpperCamelCase__ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _SCREAMING_SNAKE_CASE =os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , '''wb''' ) as fi: _SCREAMING_SNAKE_CASE =self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,)
691
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig __A : Optional[Any] = logging.get_logger(__name__) # General docstring __A : str = "PoolFormerConfig" # Base docstring __A : Optional[Any] = "sail/poolformer_s12" __A : List[Any] = [1, 512, 7, 7] # Image classification docstring __A : List[str] = "sail/poolformer_s12" __A : Tuple = "tabby, tabby cat" __A : Tuple = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def lowercase ( UpperCamelCase : Any , UpperCamelCase : float = 0.0 , UpperCamelCase : bool = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input A__ : Tuple =1 - drop_prob A__ : List[str] =(input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets A__ : Any =keep_prob + torch.rand(UpperCamelCase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize A__ : Optional[int] =input.div(UpperCamelCase ) * random_tensor return output class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : Optional[float] = None ): super().__init__() A__ : Optional[int] =drop_prob def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : torch.Tensor ): return drop_path(UpperCamelCase__ , self.drop_prob , self.training ) def _UpperCAmelCase ( self : List[str] ): return "p={}".format(self.drop_prob ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): super().__init__() A__ : Optional[int] =patch_size if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (patch_size, patch_size) A__ : Optional[int] =stride if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (stride, stride) A__ : int =padding if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (padding, padding) A__ : Any =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=UpperCamelCase__ ) A__ : Any =norm_layer(UpperCamelCase__ ) if norm_layer else nn.Identity() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : str ): A__ : List[str] =self.projection(UpperCamelCase__ ) A__ : Any =self.norm(UpperCamelCase__ ) return embeddings class __lowerCAmelCase ( nn.GroupNorm): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Dict , **UpperCamelCase__ : Union[str, Any] ): super().__init__(1 , UpperCamelCase__ , **UpperCamelCase__ ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Optional[int] ): super().__init__() A__ : Any =nn.AvgPoolad(UpperCamelCase__ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[str] ): return self.pool(UpperCamelCase__ ) - hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] ): super().__init__() A__ : List[Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Union[str, Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Dict =PoolFormerDropPath(UpperCamelCase__ ) if isinstance(config.hidden_act , UpperCamelCase__ ): A__ : Tuple =ACTaFN[config.hidden_act] else: A__ : Optional[Any] =config.hidden_act def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict ): A__ : Optional[Any] =self.conva(UpperCamelCase__ ) A__ : List[str] =self.act_fn(UpperCamelCase__ ) A__ : List[str] =self.drop(UpperCamelCase__ ) A__ : Optional[int] =self.conva(UpperCamelCase__ ) A__ : Optional[Any] =self.drop(UpperCamelCase__ ) return hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Any ): super().__init__() A__ : Optional[int] =PoolFormerPooling(UpperCamelCase__ ) A__ : List[str] =PoolFormerOutput(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) # Useful for training neural nets A__ : Tuple =PoolFormerDropPath(UpperCamelCase__ ) if drop_path > 0.0 else nn.Identity() A__ : Optional[Any] =config.use_layer_scale if config.use_layer_scale: A__ : List[str] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) A__ : List[Any] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : Optional[int] ): if self.use_layer_scale: A__ : Optional[int] =self.pooling(self.before_norm(UpperCamelCase__ ) ) A__ : Union[str, Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection A__ : Union[str, Any] =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : Tuple =() A__ : List[str] =self.output(self.after_norm(UpperCamelCase__ ) ) A__ : Optional[Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection A__ : str =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : List[Any] =(output,) + outputs return outputs else: A__ : Tuple =self.drop_path(self.pooling(self.before_norm(UpperCamelCase__ ) ) ) # First residual connection A__ : Optional[Any] =pooling_output + hidden_states A__ : Tuple =() # Second residual connection inside the PoolFormerOutput block A__ : List[str] =self.drop_path(self.output(self.after_norm(UpperCamelCase__ ) ) ) A__ : Any =hidden_states + layer_output A__ : Tuple =(output,) + outputs return outputs class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : List[str] ): super().__init__() A__ : Tuple =config # stochastic depth decay rule A__ : Dict =[x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings A__ : Tuple =[] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) A__ : List[str] =nn.ModuleList(UpperCamelCase__ ) # Transformer blocks A__ : Union[str, Any] =[] A__ : Any =0 for i in range(config.num_encoder_blocks ): # each block consists of layers A__ : Union[str, Any] =[] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCamelCase__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCamelCase__ ) ) A__ : str =nn.ModuleList(UpperCamelCase__ ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : Optional[int]=True ): A__ : Union[str, Any] =() if output_hidden_states else None A__ : Dict =pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): A__ , A__ : List[Any] =layers # Get patch embeddings from hidden_states A__ : Any =embedding_layer(UpperCamelCase__ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCamelCase__ ): A__ : List[str] =blk(UpperCamelCase__ ) A__ : Tuple =layer_outputs[0] if output_hidden_states: A__ : List[Any] =all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : List[str] = PoolFormerConfig __magic_name__ : int = """poolformer""" __magic_name__ : Any = """pixel_values""" __magic_name__ : Any = True def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : str ): if isinstance(UpperCamelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any]=False ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A__ : Optional[Any] =value __A : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : Dict ): super().__init__(UpperCamelCase__ ) A__ : List[Any] =config A__ : Optional[Any] =PoolFormerEncoder(UpperCamelCase__ ) # Initialize weights and apply final processing self.post_init() def _UpperCAmelCase ( self : Tuple ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : int =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A__ : Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) A__ : List[Any] =self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : int =encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[Any] ): super().__init__() A__ : List[str] =nn.Linear(config.hidden_size , config.hidden_size ) def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] ): A__ : int =self.dense(UpperCamelCase__ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : str ): super().__init__(UpperCamelCase__ ) A__ : List[str] =config.num_labels A__ : Optional[int] =PoolFormerModel(UpperCamelCase__ ) # Final norm A__ : Dict =PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head A__ : Dict =( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.LongTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : Tuple =return_dict if return_dict is not None else self.config.use_return_dict A__ : List[str] =self.poolformer( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : str =outputs[0] A__ : List[Any] =self.classifier(self.norm(UpperCamelCase__ ).mean([-2, -1] ) ) A__ : Optional[Any] =None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: A__ : int ="regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): A__ : Tuple ="single_label_classification" else: A__ : Optional[int] ="multi_label_classification" if self.config.problem_type == "regression": A__ : Dict =MSELoss() if self.num_labels == 1: A__ : Optional[Any] =loss_fct(logits.squeeze() , labels.squeeze() ) else: A__ : List[str] =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": A__ : Tuple =CrossEntropyLoss() A__ : int =loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": A__ : List[Any] =BCEWithLogitsLoss() A__ : str =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: A__ : Optional[int] =(logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states )
656
0
import fire from utils import calculate_rouge, save_json def a_ ( __lowercase : Optional[Any] , __lowercase : Dict , __lowercase : Dict=None , **__lowercase : List[Any] ) -> Optional[int]: _snake_case = [x.strip() for x in open(__lowercase ).readlines()] _snake_case = [x.strip() for x in open(__lowercase ).readlines()][: len(__lowercase )] _snake_case = calculate_rouge(__lowercase , __lowercase , **__lowercase ) if save_path is not None: save_json(__lowercase , __lowercase , indent=__lowercase ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
686
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : int = IFInpaintingSuperResolutionPipeline __magic_name__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __magic_name__ : str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""}) __magic_name__ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def _UpperCAmelCase ( self : Union[str, Any] ): return self._get_superresolution_dummy_components() def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[int]=0 ): if str(UpperCamelCase__ ).startswith("mps" ): A__ : Any =torch.manual_seed(UpperCamelCase__ ) else: A__ : Dict =torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A__ : Tuple =floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Optional[int] =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Any =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : List[str] ={ "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _UpperCAmelCase ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _UpperCAmelCase ( self : int ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _UpperCAmelCase ( self : Tuple ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def _UpperCAmelCase ( self : str ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _UpperCAmelCase ( self : Dict ): self._test_save_load_local() def _UpperCAmelCase ( self : Optional[int] ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
656
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCamelCase ( _UpperCamelCase ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Optional[Any] =KandinskyVaaInpaintPipeline __UpperCAmelCase : Any =["""image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] __UpperCAmelCase : List[str] =[ """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] __UpperCAmelCase : Tuple =[ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __UpperCAmelCase : List[str] =False @property def snake_case ( self ): return 32 @property def snake_case ( self ): return 32 @property def snake_case ( self ): return self.time_input_dim @property def snake_case ( self ): return self.time_input_dim * 4 @property def snake_case ( self ): return 1_00 @property def snake_case ( self ): torch.manual_seed(0 ) __lowerCAmelCase = { "in_channels": 9, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } __lowerCAmelCase = UNetaDConditionModel(**UpperCamelCase__ ) return model @property def snake_case ( self ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def snake_case ( self ): torch.manual_seed(0 ) __lowerCAmelCase = VQModel(**self.dummy_movq_kwargs ) return model def snake_case ( self ): __lowerCAmelCase = self.dummy_unet __lowerCAmelCase = self.dummy_movq __lowerCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule="linear" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=UpperCamelCase__ , set_alpha_to_one=UpperCamelCase__ , steps_offset=1 , prediction_type="epsilon" , thresholding=UpperCamelCase__ , ) __lowerCAmelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def snake_case ( self , __a , __a=0 ): __lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) __lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( UpperCamelCase__ ) # create init_image __lowerCAmelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(UpperCamelCase__ ) ).convert("RGB" ).resize((2_56, 2_56) ) # create mask __lowerCAmelCase = np.ones((64, 64) , dtype=np.floataa ) __lowerCAmelCase = 0 if str(UpperCamelCase__ ).startswith("mps" ): __lowerCAmelCase = torch.manual_seed(UpperCamelCase__ ) else: __lowerCAmelCase = torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) __lowerCAmelCase = { "image": init_image, "mask_image": mask, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 2, "guidance_scale": 4.0, "output_type": "np", } return inputs def snake_case ( self ): __lowerCAmelCase = "cpu" __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = self.pipeline_class(**UpperCamelCase__ ) __lowerCAmelCase = pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs(UpperCamelCase__ ) ) __lowerCAmelCase = output.images __lowerCAmelCase = pipe( **self.get_dummy_inputs(UpperCamelCase__ ) , return_dict=UpperCamelCase__ , )[0] __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] print(f"image.shape {image.shape}" ) assert image.shape == (1, 64, 64, 3) __lowerCAmelCase = np.array( [0.5_0_7_7_5_9_0_3, 0.4_9_5_2_7_1_9_5, 0.4_8_8_2_4_5_4_3, 0.5_0_1_9_2_2_3_7, 0.4_8_6_4_4_9_0_6, 0.4_9_3_7_3_8_1_4, 0.4_7_8_0_5_9_8, 0.4_7_2_3_4_8_2_7, 0.4_8_3_2_7_8_4_8] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" def snake_case ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case ( self ): __lowerCAmelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy" ) __lowerCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) __lowerCAmelCase = np.ones((7_68, 7_68) , dtype=np.floataa ) __lowerCAmelCase = 0 __lowerCAmelCase = "a hat" __lowerCAmelCase = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(UpperCamelCase__ ) __lowerCAmelCase = KandinskyVaaInpaintPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder-inpaint" , torch_dtype=torch.floataa ) __lowerCAmelCase = pipeline.to(UpperCamelCase__ ) pipeline.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCAmelCase = torch.Generator(device="cpu" ).manual_seed(0 ) __lowerCAmelCase = pipe_prior( UpperCamelCase__ , generator=UpperCamelCase__ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() __lowerCAmelCase = pipeline( image=UpperCamelCase__ , mask_image=UpperCamelCase__ , image_embeds=UpperCamelCase__ , negative_image_embeds=UpperCamelCase__ , generator=UpperCamelCase__ , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type="np" , ) __lowerCAmelCase = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(UpperCamelCase__ , UpperCamelCase__ )
636
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __A : Any = { "configuration_efficientformer": [ "EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientFormerConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = ["EfficientFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientFormerForImageClassification", "EfficientFormerForImageClassificationWithTeacher", "EfficientFormerModel", "EfficientFormerPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFEfficientFormerForImageClassification", "TFEfficientFormerForImageClassificationWithTeacher", "TFEfficientFormerModel", "TFEfficientFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys __A : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { "facebook/xlm-roberta-xl": "https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json", "facebook/xlm-roberta-xxl": "https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json", # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class lowerCAmelCase_ ( _UpperCamelCase ): '''simple docstring''' _lowerCamelCase: List[Any] = """xlm-roberta-xl""" def __init__( self : Any ,A_ : Optional[Any]=25_0880 ,A_ : Any=2560 ,A_ : str=36 ,A_ : List[Any]=32 ,A_ : Tuple=1_0240 ,A_ : Dict="gelu" ,A_ : List[str]=0.1 ,A_ : int=0.1 ,A_ : int=514 ,A_ : Tuple=1 ,A_ : List[Any]=0.02 ,A_ : Union[str, Any]=1e-05 ,A_ : int=1 ,A_ : Dict=0 ,A_ : List[str]=2 ,A_ : List[Any]="absolute" ,A_ : Tuple=True ,A_ : Dict=None ,**A_ : Optional[int] ,) -> Any: super().__init__(pad_token_id=UpperCamelCase__ ,bos_token_id=UpperCamelCase__ ,eos_token_id=UpperCamelCase__ ,**UpperCamelCase__ ) A = vocab_size A = hidden_size A = num_hidden_layers A = num_attention_heads A = hidden_act A = intermediate_size A = hidden_dropout_prob A = attention_probs_dropout_prob A = max_position_embeddings A = type_vocab_size A = initializer_range A = layer_norm_eps A = position_embedding_type A = use_cache A = classifier_dropout class lowerCAmelCase_ ( _UpperCamelCase ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: if self.task == "multiple-choice": A = {0: "batch", 1: "choice", 2: "sequence"} else: A = {0: "batch", 1: "sequence"} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
91
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Tuple =[] for _ in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Dict =[] for step in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] =os.path.join(UpperCamelCase , "schedule.bin" ) torch.save(scheduler.state_dict() , UpperCamelCase ) A__ : Dict =torch.load(UpperCamelCase ) scheduler.load_state_dict(UpperCamelCase ) return lrs @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : Any =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Optional[Any] =torch.tensor([0.4, 0.2, -0.5] ) A__ : Any =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : List[str] =AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(100 ): A__ : Optional[int] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def _UpperCAmelCase ( self : Dict ): A__ : Optional[int] =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Dict =torch.tensor([0.4, 0.2, -0.5] ) A__ : Optional[int] =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : int =Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=UpperCamelCase__ , weight_decay=0.0 , relative_step=UpperCamelCase__ , scale_parameter=UpperCamelCase__ , warmup_init=UpperCamelCase__ , ) for _ in range(1000 ): A__ : List[Any] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' __magic_name__ : Optional[int] = nn.Linear(50 , 50) if is_torch_available() else None __magic_name__ : Any = AdamW(m.parameters() , lr=10.0) if is_torch_available() else None __magic_name__ : Union[str, Any] = 10 def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ , msg=UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[Any] ): A__ : Union[str, Any] ={"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) A__ : Union[str, Any] ={ get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1E-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): A__ , A__ : Any =data A__ : Union[str, Any] =scheduler_func(self.optimizer , **UpperCamelCase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) A__ : int =unwrap_schedule(UpperCamelCase__ , self.num_steps ) self.assertListAlmostEqual( UpperCamelCase__ , UpperCamelCase__ , tol=1E-2 , msg=F'''failed for {scheduler_func} in normal scheduler''' , ) A__ : List[str] =scheduler_func(self.optimizer , **UpperCamelCase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(UpperCamelCase__ ) # wrap to test picklability of the schedule A__ : Tuple =unwrap_and_save_reload_schedule(UpperCamelCase__ , self.num_steps ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ , msg=F'''failed for {scheduler_func} in save and reload''' ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : str ): A__ : int =fn def __call__( self : List[Any] , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : List[Any] ): return self.fn(*UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict ): A__ : str =list(map(self , scheduler.lr_lambdas ) )
656
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
391
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __A : List[Any] = logging.get_logger("transformers.models.speecht5") __A : Optional[Any] = { "speech_encoder_prenet.layer_norm": "speecht5.encoder.prenet.feature_projection.layer_norm", "speech_encoder_prenet.post_extract_proj": "speecht5.encoder.prenet.feature_projection.projection", "speech_encoder_prenet.pos_conv.0": "speecht5.encoder.prenet.pos_conv_embed.conv", "speech_encoder_prenet.mask_emb": "speecht5.encoder.prenet.masked_spec_embed", } __A : Optional[int] = { "text_encoder_prenet.encoder_prenet.0": "speecht5.encoder.prenet.embed_tokens", "text_encoder_prenet.encoder_prenet.1.alpha": "speecht5.encoder.prenet.encode_positions.alpha", } __A : List[str] = { "speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0": "speecht5.decoder.prenet.layers.0", "speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0": "speecht5.decoder.prenet.layers.1", "speech_decoder_prenet.decoder_prenet.0.1": "speecht5.decoder.prenet.final_layer", "speech_decoder_prenet.decoder_prenet.1.alpha": "speecht5.decoder.prenet.encode_positions.alpha", "speech_decoder_prenet.spkembs_layer.0": "speecht5.decoder.prenet.speaker_embeds_layer", } __A : List[Any] = { "speech_decoder_postnet.feat_out": "speech_decoder_postnet.feat_out", "speech_decoder_postnet.prob_out": "speech_decoder_postnet.prob_out", "speech_decoder_postnet.postnet.postnet.0.0": "speech_decoder_postnet.layers.0.conv", "speech_decoder_postnet.postnet.postnet.0.1": "speech_decoder_postnet.layers.0.batch_norm", "speech_decoder_postnet.postnet.postnet.1.0": "speech_decoder_postnet.layers.1.conv", "speech_decoder_postnet.postnet.postnet.1.1": "speech_decoder_postnet.layers.1.batch_norm", "speech_decoder_postnet.postnet.postnet.2.0": "speech_decoder_postnet.layers.2.conv", "speech_decoder_postnet.postnet.postnet.2.1": "speech_decoder_postnet.layers.2.batch_norm", "speech_decoder_postnet.postnet.postnet.3.0": "speech_decoder_postnet.layers.3.conv", "speech_decoder_postnet.postnet.postnet.3.1": "speech_decoder_postnet.layers.3.batch_norm", "speech_decoder_postnet.postnet.postnet.4.0": "speech_decoder_postnet.layers.4.conv", "speech_decoder_postnet.postnet.postnet.4.1": "speech_decoder_postnet.layers.4.batch_norm", } __A : Union[str, Any] = { "text_decoder_prenet.embed_tokens": "speecht5.decoder.prenet.embed_tokens", } __A : Any = { "text_decoder_postnet.output_projection": "text_decoder_postnet.lm_head", } __A : Union[str, Any] = { "encoder.layers.*.self_attn.k_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj", "encoder.layers.*.self_attn.v_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj", "encoder.layers.*.self_attn.q_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj", "encoder.layers.*.self_attn.out_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj", "encoder.layers.*.self_attn_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.layer_norm", "encoder.layers.*.fc1": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense", "encoder.layers.*.fc2": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense", "encoder.layers.*.final_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm", "encoder.layer_norm": "speecht5.encoder.wrapped_encoder.layer_norm", "encoder.pos_emb.pe_k": "speecht5.encoder.wrapped_encoder.embed_positions.pe_k", } __A : Optional[int] = { "decoder.layers.*.self_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj", "decoder.layers.*.self_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj", "decoder.layers.*.self_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj", "decoder.layers.*.self_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj", "decoder.layers.*.self_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm", "decoder.layers.*.encoder_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj", "decoder.layers.*.encoder_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj", "decoder.layers.*.encoder_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj", "decoder.layers.*.encoder_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj", "decoder.layers.*.encoder_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm", "decoder.layers.*.fc1": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense", "decoder.layers.*.fc2": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense", "decoder.layers.*.final_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm", } __A : Union[str, Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __A : Optional[Any] = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : Optional[int] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : int = [] __A : int = [ "encoder.version", "encoder.layers.*.norm_k.weight", "encoder.layers.*.norm_k.bias", "decoder.version", "decoder.layers.*.norm_k.weight", "decoder.layers.*.norm_k.bias", "decoder.pos_emb.pe_k", "speech_encoder_prenet.embed_positions._float_tensor", "text_decoder_prenet.embed_positions._float_tensor", ] __A : Optional[Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "speech_decoder_prenet.*", "speech_decoder_postnet.*", ] __A : Tuple = IGNORE_KEYS + [ "encoder.proj", "speech_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] __A : Union[str, Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] def lowercase ( UpperCamelCase : str , UpperCamelCase : Union[str, Any] , UpperCamelCase : int , UpperCamelCase : List[Any] , UpperCamelCase : int ): """simple docstring""" for attribute in key.split("." ): A__ : Dict =getattr(UpperCamelCase , UpperCamelCase ) if weight_type is not None: A__ : Union[str, Any] =getattr(UpperCamelCase , UpperCamelCase ).shape else: A__ : Tuple =hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": A__ : Any =value elif weight_type == "weight_g": A__ : Any =value elif weight_type == "weight_v": A__ : Any =value elif weight_type == "bias": A__ : Tuple =value elif weight_type == "running_mean": A__ : Dict =value elif weight_type == "running_var": A__ : List[str] =value elif weight_type == "num_batches_tracked": A__ : Dict =value else: A__ : Optional[int] =value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''' ) def lowercase ( UpperCamelCase : Tuple , UpperCamelCase : Tuple ): """simple docstring""" for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: A__ , A__ : List[str] =key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : Dict ): """simple docstring""" A__ : Tuple =[] if task == "s2t": A__ : Dict =hf_model.speechta.encoder.prenet.feature_encoder A__ : int =MAPPING_S2T A__ : List[Any] =IGNORE_KEYS_S2T elif task == "t2s": A__ : Union[str, Any] =None A__ : List[Any] =MAPPING_T2S A__ : Tuple =IGNORE_KEYS_T2S elif task == "s2s": A__ : Optional[Any] =hf_model.speechta.encoder.prenet.feature_encoder A__ : Tuple =MAPPING_S2S A__ : Any =IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''' ) for name, value in fairseq_dict.items(): if should_ignore(UpperCamelCase , UpperCamelCase ): logger.info(F'''{name} was ignored''' ) continue A__ : Optional[Any] =False if "conv_layers" in name: load_conv_layer( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , hf_model.config.feat_extract_norm == "group" , ) A__ : List[Any] =True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: A__ , A__ : Dict =key.split(".*." ) if prefix in name and suffix in name: A__ : int =suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: A__ : List[Any] =True if "*" in mapped_key: A__ : Optional[int] =name.split(UpperCamelCase )[0].split("." )[-2] A__ : int =mapped_key.replace("*" , UpperCamelCase ) if "weight_g" in name: A__ : str ="weight_g" elif "weight_v" in name: A__ : Optional[Any] ="weight_v" elif "bias" in name: A__ : Any ="bias" elif "weight" in name: A__ : Optional[int] ="weight" elif "running_mean" in name: A__ : Tuple ="running_mean" elif "running_var" in name: A__ : Optional[int] ="running_var" elif "num_batches_tracked" in name: A__ : str ="num_batches_tracked" else: A__ : List[Any] =None set_recursively(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) continue if not is_used: unused_weights.append(UpperCamelCase ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Dict , UpperCamelCase : Any , UpperCamelCase : str , UpperCamelCase : Dict ): """simple docstring""" A__ : Any =full_name.split("conv_layers." )[-1] A__ : Dict =name.split("." ) A__ : int =int(items[0] ) A__ : str =int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) A__ : Optional[Any] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) A__ : Optional[int] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(UpperCamelCase ) @torch.no_grad() def lowercase ( UpperCamelCase : Any , UpperCamelCase : Union[str, Any] , UpperCamelCase : List[str] , UpperCamelCase : str=None , UpperCamelCase : Any=None , UpperCamelCase : Tuple=None , ): """simple docstring""" if config_path is not None: A__ : Any =SpeechTaConfig.from_pretrained(UpperCamelCase ) else: A__ : Any =SpeechTaConfig() if task == "s2t": A__ : Union[str, Any] =config.max_text_positions A__ : Dict =SpeechTaForSpeechToText(UpperCamelCase ) elif task == "t2s": A__ : str =1876 A__ : Optional[int] =600 A__ : Tuple =config.max_speech_positions A__ : Optional[Any] =SpeechTaForTextToSpeech(UpperCamelCase ) elif task == "s2s": A__ : str =1876 A__ : Tuple =config.max_speech_positions A__ : Any =SpeechTaForSpeechToSpeech(UpperCamelCase ) else: raise ValueError(F'''Unknown task name: {task}''' ) if vocab_path: A__ : str =SpeechTaTokenizer(UpperCamelCase , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it A__ : Optional[Any] =AddedToken("<mask>" , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) A__ : int =mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) A__ : Dict =SpeechTaFeatureExtractor() A__ : Tuple =SpeechTaProcessor(tokenizer=UpperCamelCase , feature_extractor=UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) A__ : Union[str, Any] =torch.load(UpperCamelCase ) recursively_load_weights(fairseq_checkpoint["model"] , UpperCamelCase , UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if repo_id: print("Pushing to the hub..." ) processor.push_to_hub(UpperCamelCase ) model.push_to_hub(UpperCamelCase ) if __name__ == "__main__": __A : Dict = argparse.ArgumentParser() parser.add_argument( "--task", default="s2t", type=str, help="Type of the SpeechT5 model you'd like to convert. Should be one of 's2t', 't2s', 's2s'.", ) parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--vocab_path", default=None, type=str, help="Path to SentencePiece model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) __A : str = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
656
0
'''simple docstring''' def a_ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): def count_of_possible_combinations(UpperCamelCase_ ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase_ ) def a_ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): def count_of_possible_combinations_with_dp_array( UpperCamelCase_ , UpperCamelCase_ ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] A_ = sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase_ ) for item in array ) A_ = answer return answer A_ = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase_ , UpperCamelCase_ ) def a_ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): A_ = [0] * (target + 1) A_ = 1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase_ ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : Optional[Any] = 3 __SCREAMING_SNAKE_CASE : Optional[Any] = 5 __SCREAMING_SNAKE_CASE : int = [1, 2, 5] print(combination_sum_iv(n, array, target))
452
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase): '''simple docstring''' __magic_name__ : List[Any] = [R"""h\.\d+\.attn\.bias""", R"""h\.\d+\.attn\.masked_bias"""] @register_to_config def __init__( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : int = 50257 , UpperCamelCase__ : int = 1024 , UpperCamelCase__ : int = 768 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : str = "gelu_new" , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 1E-5 , UpperCamelCase__ : float = 0.02 , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , ): super().__init__() A__ : Dict =prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) A__ : Optional[int] =prefix_inner_dim A__ : Optional[int] =prefix_hidden_dim A__ : Optional[int] =( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : Optional[int] =( nn.Linear(self.prefix_hidden_dim , UpperCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : str =GPTaConfig( vocab_size=UpperCamelCase__ , n_positions=UpperCamelCase__ , n_embd=UpperCamelCase__ , n_layer=UpperCamelCase__ , n_head=UpperCamelCase__ , n_inner=UpperCamelCase__ , activation_function=UpperCamelCase__ , resid_pdrop=UpperCamelCase__ , embd_pdrop=UpperCamelCase__ , attn_pdrop=UpperCamelCase__ , layer_norm_epsilon=UpperCamelCase__ , initializer_range=UpperCamelCase__ , scale_attn_weights=UpperCamelCase__ , use_cache=UpperCamelCase__ , scale_attn_by_inverse_layer_idx=UpperCamelCase__ , reorder_and_upcast_attn=UpperCamelCase__ , ) A__ : Any =GPTaLMHeadModel(UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : Optional[torch.Tensor] = None , UpperCamelCase__ : Optional[torch.Tensor] = None , ): A__ : int =self.transformer.transformer.wte(UpperCamelCase__ ) A__ : Tuple =self.encode_prefix(UpperCamelCase__ ) A__ : Union[str, Any] =self.decode_prefix(UpperCamelCase__ ) A__ : Tuple =torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A__ : Any =self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A__ : List[Any] =torch.cat((dummy_token, input_ids) , dim=1 ) A__ : Any =self.transformer(inputs_embeds=UpperCamelCase__ , labels=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : torch.device ): return torch.zeros(UpperCamelCase__ , self.prefix_length , dtype=torch.intaa , device=UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Tuple ): return self.encode_prefix(UpperCamelCase__ ) @torch.no_grad() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : str ): A__ : Optional[int] =torch.split(UpperCamelCase__ , 1 , dim=0 ) A__ : List[str] =[] A__ : Dict =[] for feature in features: A__ : Any =self.decode_prefix(feature.to(UpperCamelCase__ ) ) # back to the clip feature # Only support beam search for now A__ , A__ : Optional[Any] =self.generate_beam( input_embeds=UpperCamelCase__ , device=UpperCamelCase__ , eos_token_id=UpperCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A__ : Optional[Any] =torch.stack(UpperCamelCase__ ) A__ : Optional[int] =torch.stack(UpperCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : List[str]=None , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : int = 5 , UpperCamelCase__ : int = 67 , UpperCamelCase__ : float = 1.0 , UpperCamelCase__ : Optional[int] = None , ): A__ : str =eos_token_id A__ : Optional[Any] =None A__ : int =None A__ : Union[str, Any] =torch.ones(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.int ) A__ : Any =torch.zeros(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.bool ) if input_embeds is not None: A__ : Union[str, Any] =input_embeds else: A__ : Optional[Any] =self.transformer.transformer.wte(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): A__ : Optional[int] =self.transformer(inputs_embeds=UpperCamelCase__ ) A__ : Tuple =outputs.logits A__ : Union[str, Any] =logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A__ : Optional[Any] =logits.softmax(-1 ).log() if scores is None: A__ , A__ : Union[str, Any] =logits.topk(UpperCamelCase__ , -1 ) A__ : Union[str, Any] =generated.expand(UpperCamelCase__ , *generated.shape[1:] ) A__ , A__ : Optional[int] =next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A__ : str =next_tokens else: A__ : Optional[Any] =tokens.expand(UpperCamelCase__ , *tokens.shape[1:] ) A__ : str =torch.cat((tokens, next_tokens) , dim=1 ) else: A__ : Union[str, Any] =-float(np.inf ) A__ : Dict =0 A__ : Optional[Any] =scores[:, None] + logits seq_lengths[~is_stopped] += 1 A__ : Optional[Any] =scores_sum / seq_lengths[:, None] A__ , A__ : List[Any] =scores_sum_average.view(-1 ).topk(UpperCamelCase__ , -1 ) A__ : Tuple =next_tokens // scores_sum.shape[1] A__ : List[Any] =seq_lengths[next_tokens_source] A__ : int =next_tokens % scores_sum.shape[1] A__ : str =next_tokens.unsqueeze(1 ) A__ : List[Any] =tokens[next_tokens_source] A__ : int =torch.cat((tokens, next_tokens) , dim=1 ) A__ : List[str] =generated[next_tokens_source] A__ : Optional[Any] =scores_sum_average * seq_lengths A__ : Optional[int] =is_stopped[next_tokens_source] A__ : List[str] =self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A__ : str =torch.cat((generated, next_token_embed) , dim=1 ) A__ : str =is_stopped + next_tokens.eq(UpperCamelCase__ ).squeeze() if is_stopped.all(): break A__ : Optional[int] =scores / seq_lengths A__ : List[Any] =scores.argsort(descending=UpperCamelCase__ ) # tokens tensors are already padded to max_seq_length A__ : int =[tokens[i] for i in order] A__ : Any =torch.stack(UpperCamelCase__ , dim=0 ) A__ : int =torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
656
0
"""simple docstring""" import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class __snake_case (_UpperCamelCase ): __a = """MCTCTFeatureExtractor""" __a = """AutoTokenizer""" def __init__( self: Optional[Any] , A_: int , A_: Optional[int] ): super().__init__(UpperCamelCase__ , UpperCamelCase__ ) __lowerCamelCase = self.feature_extractor __lowerCamelCase = False def __call__( self: str , *A_: Dict , **A_: Dict ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*UpperCamelCase__ , **UpperCamelCase__ ) if "raw_speech" in kwargs: warnings.warn("""Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.""" ) __lowerCamelCase = kwargs.pop("""raw_speech""" ) else: __lowerCamelCase = kwargs.pop("""audio""" , UpperCamelCase__ ) __lowerCamelCase = kwargs.pop("""sampling_rate""" , UpperCamelCase__ ) __lowerCamelCase = kwargs.pop("""text""" , UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: __lowerCamelCase = args[0] __lowerCamelCase = args[1:] if audio is None and text is None: raise ValueError("""You need to specify either an `audio` or `text` input to process.""" ) if audio is not None: __lowerCamelCase = self.feature_extractor(UpperCamelCase__ , *UpperCamelCase__ , sampling_rate=UpperCamelCase__ , **UpperCamelCase__ ) if text is not None: __lowerCamelCase = self.tokenizer(UpperCamelCase__ , **UpperCamelCase__ ) if text is None: return inputs elif audio is None: return encodings else: __lowerCamelCase = encodings["input_ids"] return inputs def __a ( self: str , *A_: Union[str, Any] , **A_: Optional[int] ): return self.tokenizer.batch_decode(*UpperCamelCase__ , **UpperCamelCase__ ) def __a ( self: Optional[Any] , *A_: Optional[Any] , **A_: List[str] ): # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*UpperCamelCase__ , **UpperCamelCase__ ) __lowerCamelCase = kwargs.pop("""input_features""" , UpperCamelCase__ ) __lowerCamelCase = kwargs.pop("""labels""" , UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: __lowerCamelCase = args[0] __lowerCamelCase = args[1:] if input_features is not None: __lowerCamelCase = self.feature_extractor.pad(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) if labels is not None: __lowerCamelCase = self.tokenizer.pad(UpperCamelCase__ , **UpperCamelCase__ ) if labels is None: return input_features elif input_features is None: return labels else: __lowerCamelCase = labels["input_ids"] return input_features def __a ( self: Any , *A_: List[Any] , **A_: Tuple ): return self.tokenizer.decode(*UpperCamelCase__ , **UpperCamelCase__ ) @contextmanager def __a ( self: Optional[Any] ): warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your audio inputs, or in a separate call.""" ) __lowerCamelCase = True __lowerCamelCase = self.tokenizer yield __lowerCamelCase = self.feature_extractor __lowerCamelCase = False
281
"""simple docstring""" import os def lowercase ( ): """simple docstring""" A__ : List[Any] =os.path.dirname(os.path.realpath(UpperCamelCase ) ) A__ : str =os.path.join(UpperCamelCase , "triangle.txt" ) with open(UpperCamelCase ) as f: A__ : Optional[int] =f.readlines() A__ : str =[] for line in triangle: A__ : Union[str, Any] =[] for number in line.strip().split(" " ): numbers_from_line.append(int(UpperCamelCase ) ) a.append(UpperCamelCase ) for i in range(1 , len(UpperCamelCase ) ): for j in range(len(a[i] ) ): A__ : Union[str, Any] =a[i - 1][j] if j != len(a[i - 1] ) else 0 A__ : Union[str, Any] =a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(UpperCamelCase , UpperCamelCase ) return max(a[-1] ) if __name__ == "__main__": print(solution())
656
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer _UpperCamelCase : List[str] =logging.get_logger(__name__) _UpperCamelCase : List[str] ={"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _UpperCamelCase : Dict ={ "vocab_file": { "bert-base-uncased": "https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt", "bert-large-uncased": "https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt", "bert-base-cased": "https://huggingface.co/bert-base-cased/resolve/main/vocab.txt", "bert-large-cased": "https://huggingface.co/bert-large-cased/resolve/main/vocab.txt", "bert-base-multilingual-uncased": ( "https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt" ), "bert-base-multilingual-cased": "https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt", "bert-base-chinese": "https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt", "bert-base-german-cased": "https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt", "bert-large-uncased-whole-word-masking": ( "https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt" ), "bert-large-cased-whole-word-masking": ( "https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt" ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt" ), "bert-large-cased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt" ), "bert-base-cased-finetuned-mrpc": ( "https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt" ), "bert-base-german-dbmdz-cased": "https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt", "bert-base-german-dbmdz-uncased": ( "https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt" ), "TurkuNLP/bert-base-finnish-cased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt" ), "TurkuNLP/bert-base-finnish-uncased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt" ), "wietsedv/bert-base-dutch-cased": ( "https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "bert-base-uncased": "https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json", "bert-large-uncased": "https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json", "bert-base-cased": "https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json", "bert-large-cased": "https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json", "bert-base-multilingual-uncased": ( "https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json" ), "bert-base-multilingual-cased": ( "https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json" ), "bert-base-chinese": "https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json", "bert-base-german-cased": "https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json", "bert-large-uncased-whole-word-masking": ( "https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json" ), "bert-large-cased-whole-word-masking": ( "https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json" ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json" ), "bert-large-cased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json" ), "bert-base-cased-finetuned-mrpc": ( "https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json" ), "bert-base-german-dbmdz-cased": ( "https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json" ), "bert-base-german-dbmdz-uncased": ( "https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json" ), "TurkuNLP/bert-base-finnish-cased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json" ), "TurkuNLP/bert-base-finnish-uncased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json" ), "wietsedv/bert-base-dutch-cased": ( "https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json" ), }, } _UpperCamelCase : int ={ "bert-base-uncased": 512, "bert-large-uncased": 512, "bert-base-cased": 512, "bert-large-cased": 512, "bert-base-multilingual-uncased": 512, "bert-base-multilingual-cased": 512, "bert-base-chinese": 512, "bert-base-german-cased": 512, "bert-large-uncased-whole-word-masking": 512, "bert-large-cased-whole-word-masking": 512, "bert-large-uncased-whole-word-masking-finetuned-squad": 512, "bert-large-cased-whole-word-masking-finetuned-squad": 512, "bert-base-cased-finetuned-mrpc": 512, "bert-base-german-dbmdz-cased": 512, "bert-base-german-dbmdz-uncased": 512, "TurkuNLP/bert-base-finnish-cased-v1": 512, "TurkuNLP/bert-base-finnish-uncased-v1": 512, "wietsedv/bert-base-dutch-cased": 512, } _UpperCamelCase : Optional[int] ={ "bert-base-uncased": {"do_lower_case": True}, "bert-large-uncased": {"do_lower_case": True}, "bert-base-cased": {"do_lower_case": False}, "bert-large-cased": {"do_lower_case": False}, "bert-base-multilingual-uncased": {"do_lower_case": True}, "bert-base-multilingual-cased": {"do_lower_case": False}, "bert-base-chinese": {"do_lower_case": False}, "bert-base-german-cased": {"do_lower_case": False}, "bert-large-uncased-whole-word-masking": {"do_lower_case": True}, "bert-large-cased-whole-word-masking": {"do_lower_case": False}, "bert-large-uncased-whole-word-masking-finetuned-squad": {"do_lower_case": True}, "bert-large-cased-whole-word-masking-finetuned-squad": {"do_lower_case": False}, "bert-base-cased-finetuned-mrpc": {"do_lower_case": False}, "bert-base-german-dbmdz-cased": {"do_lower_case": False}, "bert-base-german-dbmdz-uncased": {"do_lower_case": True}, "TurkuNLP/bert-base-finnish-cased-v1": {"do_lower_case": False}, "TurkuNLP/bert-base-finnish-uncased-v1": {"do_lower_case": True}, "wietsedv/bert-base-dutch-cased": {"do_lower_case": False}, } class UpperCAmelCase__ ( _UpperCamelCase ): __snake_case : List[str] = VOCAB_FILES_NAMES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : int = PRETRAINED_INIT_CONFIGURATION __snake_case : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : Optional[int] = BertTokenizer def __init__( self ,A__=None ,A__=None ,A__=True ,A__="[UNK]" ,A__="[SEP]" ,A__="[PAD]" ,A__="[CLS]" ,A__="[MASK]" ,A__=True ,A__=None ,**A__ ,): super().__init__( UpperCamelCase__ ,tokenizer_file=UpperCamelCase__ ,do_lower_case=UpperCamelCase__ ,unk_token=UpperCamelCase__ ,sep_token=UpperCamelCase__ ,pad_token=UpperCamelCase__ ,cls_token=UpperCamelCase__ ,mask_token=UpperCamelCase__ ,tokenize_chinese_chars=UpperCamelCase__ ,strip_accents=UpperCamelCase__ ,**UpperCamelCase__ ,) _A : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' ,UpperCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' ,UpperCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' ,UpperCamelCase__ ) != tokenize_chinese_chars ): _A : Union[str, Any] = getattr(UpperCamelCase__ ,normalizer_state.pop('''type''' ) ) _A : Dict = do_lower_case _A : Dict = strip_accents _A : Optional[int] = tokenize_chinese_chars _A : List[str] = normalizer_class(**UpperCamelCase__ ) _A : Optional[Any] = do_lower_case def A__ ( self ,A__ ,A__=None ): _A : int = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self ,A__ ,A__ = None ): _A : Tuple = [self.sep_token_id] _A : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self ,A__ ,A__ = None ): _A : Any = self._tokenizer.model.save(UpperCamelCase__ ,name=UpperCamelCase__ ) return tuple(UpperCamelCase__ )
206
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A : int = logging.get_logger(__name__) def lowercase ( UpperCamelCase : Any ): """simple docstring""" A__ : str =OrderedDict() for key, value in state_dict.items(): if key.startswith("module.encoder" ): A__ : Dict =key.replace("module.encoder" , "glpn.encoder" ) if key.startswith("module.decoder" ): A__ : Optional[int] =key.replace("module.decoder" , "decoder.stages" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 A__ : Tuple =key[key.find("patch_embed" ) + len("patch_embed" )] A__ : Optional[Any] =key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(UpperCamelCase )-1}''' ) if "norm" in key: A__ : Dict =key.replace("norm" , "layer_norm" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 A__ : Any =key[key.find("glpn.encoder.layer_norm" ) + len("glpn.encoder.layer_norm" )] A__ : Tuple =key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(UpperCamelCase )-1}''' ) if "layer_norm1" in key: A__ : List[Any] =key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: A__ : Optional[int] =key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 A__ : int =key[key.find("block" ) + len("block" )] A__ : Optional[Any] =key.replace(F'''block{idx}''' , F'''block.{int(UpperCamelCase )-1}''' ) if "attn.q" in key: A__ : Optional[Any] =key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: A__ : Union[str, Any] =key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: A__ : str =key.replace("attn" , "attention.self" ) if "fc1" in key: A__ : Dict =key.replace("fc1" , "dense1" ) if "fc2" in key: A__ : str =key.replace("fc2" , "dense2" ) if "linear_pred" in key: A__ : List[Any] =key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: A__ : List[str] =key.replace("linear_fuse.conv" , "linear_fuse" ) A__ : Any =key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 A__ : str =key[key.find("linear_c" ) + len("linear_c" )] A__ : Dict =key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(UpperCamelCase )-1}''' ) if "bot_conv" in key: A__ : Union[str, Any] =key.replace("bot_conv" , "0.convolution" ) if "skip_conv1" in key: A__ : List[Any] =key.replace("skip_conv1" , "1.convolution" ) if "skip_conv2" in key: A__ : int =key.replace("skip_conv2" , "2.convolution" ) if "fusion1" in key: A__ : Optional[Any] =key.replace("fusion1" , "1.fusion" ) if "fusion2" in key: A__ : Optional[Any] =key.replace("fusion2" , "2.fusion" ) if "fusion3" in key: A__ : int =key.replace("fusion3" , "3.fusion" ) if "fusion" in key and "conv" in key: A__ : List[str] =key.replace("conv" , "convolutional_layer" ) if key.startswith("module.last_layer_depth" ): A__ : Tuple =key.replace("module.last_layer_depth" , "head.head" ) A__ : int =value return new_state_dict def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Dict ): """simple docstring""" # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) A__ : int =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) A__ : str =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict A__ : List[str] =kv_weight[ : config.hidden_sizes[i], : ] A__ : Dict =kv_bias[: config.hidden_sizes[i]] A__ : Any =kv_weight[ config.hidden_sizes[i] :, : ] A__ : Any =kv_bias[config.hidden_sizes[i] :] def lowercase ( ): """simple docstring""" A__ : Optional[Any] ="http://images.cocodataset.org/val2017/000000039769.jpg" A__ : List[Any] =Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return image @torch.no_grad() def lowercase ( UpperCamelCase : str , UpperCamelCase : Tuple , UpperCamelCase : List[str]=False , UpperCamelCase : str=None ): """simple docstring""" A__ : List[str] =GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) A__ : str =GLPNImageProcessor() # prepare image A__ : Any =prepare_img() A__ : Optional[int] =image_processor(images=UpperCamelCase , return_tensors="pt" ).pixel_values logger.info("Converting model..." ) # load original state dict A__ : int =torch.load(UpperCamelCase , map_location=torch.device("cpu" ) ) # rename keys A__ : Union[str, Any] =rename_keys(UpperCamelCase ) # key and value matrices need special treatment read_in_k_v(UpperCamelCase , UpperCamelCase ) # create HuggingFace model and load state dict A__ : Optional[int] =GLPNForDepthEstimation(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() # forward pass A__ : int =model(UpperCamelCase ) A__ : Optional[Any] =outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: A__ : List[Any] =torch.tensor( [[4.41_47, 4.08_73, 4.06_73], [3.78_90, 3.28_81, 3.15_25], [3.76_74, 3.54_23, 3.49_13]] ) elif "kitti" in model_name: A__ : Tuple =torch.tensor( [[3.42_91, 2.78_65, 2.51_51], [3.28_41, 2.70_21, 2.35_02], [3.11_47, 2.46_25, 2.24_81]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) A__ : str =torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , UpperCamelCase , atol=1E-4 ) print("Looks ok!" ) # finally, push to hub if required if push_to_hub: logger.info("Pushing model and image processor to the hub..." ) model.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=UpperCamelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=UpperCamelCase , ) if __name__ == "__main__": __A : List[str] = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) parser.add_argument( "--model_name", default="glpn-kitti", type=str, help="Name of the model in case you're pushing to the hub.", ) __A : Any = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
656
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = {"vocab_file": "spiece.model"} _SCREAMING_SNAKE_CASE = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class __magic_name__ ( _UpperCamelCase ): def __init__( self : List[str] , snake_case_ : List[Any] , snake_case_ : Union[str, Any]=False , snake_case_ : Dict=True , snake_case_ : List[Any]=False , snake_case_ : Dict="<s>" , snake_case_ : str="</s>" , snake_case_ : Union[str, Any]="<unk>" , snake_case_ : Optional[int]="<sep>" , snake_case_ : Optional[int]="<pad>" , snake_case_ : Optional[int]="<cls>" , snake_case_ : List[str]="<mask>" , snake_case_ : Optional[Any]=["<eop>", "<eod>"] , snake_case_ : Optional[Dict[str, Any]] = None , **snake_case_ : Dict , ): __snake_case = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token __snake_case = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) __snake_case = 3 __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) __snake_case = jieba __snake_case = str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def lowerCAmelCase ( self : Union[str, Any] ): return len(self.sp_model ) def lowerCAmelCase ( self : Optional[int] ): __snake_case = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[str] ): __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : Tuple , snake_case_ : int ): __snake_case = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __snake_case = {} __snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCAmelCase ( self : Union[str, Any] , snake_case_ : Dict ): if self.remove_space: __snake_case = " ".join(inputs.strip().split() ) else: __snake_case = inputs __snake_case = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: __snake_case = unicodedata.normalize("NFKD" , UpperCamelCase__ ) __snake_case = "".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: __snake_case = outputs.lower() return outputs def lowerCAmelCase ( self : Optional[int] , snake_case_ : str ): __snake_case = self.preprocess_text(UpperCamelCase__ ) __snake_case = self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) __snake_case = [] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __snake_case = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __snake_case = cur_pieces[1:] else: __snake_case = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def lowerCAmelCase ( self : int , snake_case_ : str ): return self.sp_model.PieceToId(UpperCamelCase__ ) def lowerCAmelCase ( self : List[str] , snake_case_ : List[Any] ): return self.sp_model.IdToPiece(UpperCamelCase__ ) def lowerCAmelCase ( self : Union[str, Any] , snake_case_ : str ): __snake_case = "".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def lowerCAmelCase ( self : Optional[int] , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowerCAmelCase ( self : Optional[int] , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None , snake_case_ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def lowerCAmelCase ( self : int , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): __snake_case = [self.sep_token_id] __snake_case = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowerCAmelCase ( self : Dict , snake_case_ : str , snake_case_ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return __snake_case = os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: __snake_case = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def lowerCAmelCase ( self : str , *snake_case_ : Optional[Any] , **snake_case_ : int ): __snake_case = super()._decode(*UpperCamelCase__ , **UpperCamelCase__ ) __snake_case = text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
163
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging __A : Any = logging.get_logger(__name__) __A : Optional[Any] = { "EleutherAI/gpt-neo-1.3B": "https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Union[str, Any] = """gpt_neo""" __magic_name__ : Union[str, Any] = ["""past_key_values"""] __magic_name__ : Dict = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self : Dict , UpperCamelCase__ : List[Any]=50257 , UpperCamelCase__ : Optional[Any]=2048 , UpperCamelCase__ : Tuple=2048 , UpperCamelCase__ : int=24 , UpperCamelCase__ : Dict=[[["global", "local"], 12]] , UpperCamelCase__ : Optional[Any]=16 , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : str=256 , UpperCamelCase__ : List[str]="gelu_new" , UpperCamelCase__ : List[str]=0.0 , UpperCamelCase__ : Tuple=0.0 , UpperCamelCase__ : str=0.0 , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : List[str]=1E-5 , UpperCamelCase__ : Any=0.02 , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : Optional[Any]=50256 , UpperCamelCase__ : List[str]=50256 , **UpperCamelCase__ : str , ): A__ : Optional[Any] =vocab_size A__ : Dict =max_position_embeddings A__ : List[str] =hidden_size A__ : List[Any] =num_layers A__ : Tuple =num_heads A__ : List[str] =intermediate_size A__ : Tuple =window_size A__ : Dict =activation_function A__ : str =resid_dropout A__ : Union[str, Any] =embed_dropout A__ : List[str] =attention_dropout A__ : Tuple =classifier_dropout A__ : int =layer_norm_epsilon A__ : int =initializer_range A__ : str =use_cache A__ : Tuple =bos_token_id A__ : int =eos_token_id A__ : int =attention_types A__ : Any =self.expand_attention_types_params(UpperCamelCase__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.attention_layers)` == `config.num_layers` " F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' "`config.attention_layers` is prepared using `config.attention_types`. " "Please verify the value of `config.attention_types` argument." ) super().__init__(bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) @staticmethod def _UpperCAmelCase ( UpperCamelCase__ : List[str] ): A__ : Optional[Any] =[] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Optional[Any] , UpperCamelCase : List[Any] , UpperCamelCase : Optional[int] ): """simple docstring""" import torch A__ : List[str] =input.size() A__ : Dict =len(UpperCamelCase ) A__ : Optional[int] =shape[dimension] A__ : str =torch.arange(0 , UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =torch.div(sizedim - size , UpperCamelCase , rounding_mode="floor" ) + 1 A__ : str =torch.arange(UpperCamelCase ) + low_indices[:min_length][:, None] A__ : Tuple =[slice(UpperCamelCase )] * rank A__ : int =indices A__ : Optional[int] =input[s] A__ : Union[str, Any] =list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(UpperCamelCase ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any ): """simple docstring""" import torch A__ : List[str] =torch.arange(1 , UpperCamelCase ) A__ : List[Any] =torch.remainder(UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =remainders == 0 A__ : str =candidates[divisor_indices] A__ : int =torch.max(UpperCamelCase ) return largest_divisor, torch.div(UpperCamelCase , UpperCamelCase , rounding_mode="floor" ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' @property def _UpperCAmelCase ( self : List[Any] ): A__ : Optional[int] =OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(UpperCamelCase__ , direction="inputs" ) A__ : Optional[int] ={0: "batch", 1: "past_sequence + sequence"} else: A__ : Tuple ={0: "batch", 1: "sequence"} return common_inputs @property def _UpperCAmelCase ( self : List[str] ): return self._config.num_heads def _UpperCAmelCase ( self : int , UpperCamelCase__ : PreTrainedTokenizer , UpperCamelCase__ : int = -1 , UpperCamelCase__ : int = -1 , UpperCamelCase__ : bool = False , UpperCamelCase__ : Optional[TensorType] = None , ): A__ : Union[str, Any] =super(UpperCamelCase__ , self ).generate_dummy_inputs( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) # We need to order the input in the way they appears in the forward() A__ : List[Any] =OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A__ , A__ : Union[str, Any] =common_inputs["input_ids"].shape # Not using the same length for past_key_values A__ : Union[str, Any] =seqlen + 2 A__ : List[Any] =( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) A__ : Optional[Any] =[ (torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) for _ in range(self.num_layers ) ] A__ : Optional[Any] =common_inputs["attention_mask"] if self.use_past: A__ : Any =ordered_inputs["attention_mask"].dtype A__ : Tuple =torch.cat( [ordered_inputs["attention_mask"], torch.ones(UpperCamelCase__ , UpperCamelCase__ , dtype=UpperCamelCase__ )] , dim=1 ) return ordered_inputs @property def _UpperCAmelCase ( self : List[str] ): return 13
656
0
"""simple docstring""" def UpperCamelCase ( ) ->List[Any]: return 1 def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->Optional[int]: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->Union[str, Any]: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->Dict: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->List[Any]: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->Optional[Any]: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ ) ->Dict: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ = 200 ) ->List[str]: return two_pound(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": print(solution(int(input().strip())))
434
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Union[str, Any] = logging.get_logger(__name__) __A : Any = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Tuple = """megatron-bert""" def __init__( self : Tuple , UpperCamelCase__ : Dict=29056 , UpperCamelCase__ : int=1024 , UpperCamelCase__ : Optional[int]=24 , UpperCamelCase__ : Dict=16 , UpperCamelCase__ : int=4096 , UpperCamelCase__ : str="gelu" , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : int=0.1 , UpperCamelCase__ : int=512 , UpperCamelCase__ : str=2 , UpperCamelCase__ : Union[str, Any]=0.02 , UpperCamelCase__ : Any=1E-12 , UpperCamelCase__ : List[Any]=0 , UpperCamelCase__ : str="absolute" , UpperCamelCase__ : Dict=True , **UpperCamelCase__ : Tuple , ): super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) A__ : Optional[int] =vocab_size A__ : Optional[int] =hidden_size A__ : str =num_hidden_layers A__ : Any =num_attention_heads A__ : str =hidden_act A__ : Optional[int] =intermediate_size A__ : str =hidden_dropout_prob A__ : str =attention_probs_dropout_prob A__ : List[Any] =max_position_embeddings A__ : List[Any] =type_vocab_size A__ : Tuple =initializer_range A__ : Any =layer_norm_eps A__ : Any =position_embedding_type A__ : Union[str, Any] =use_cache
656
0
from __future__ import annotations from random import choice def __a ( A__ : Tuple ): return choice(A__ ) def __a ( A__ : list[int] , A__ : int ): SCREAMING_SNAKE_CASE = random_pivot(A__ ) # partition based on pivot # linear time SCREAMING_SNAKE_CASE = [e for e in lst if e < pivot] SCREAMING_SNAKE_CASE = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(A__ ) == k - 1: return pivot # pivot is in elements bigger than k elif len(A__ ) < k - 1: return kth_number(A__ , k - len(A__ ) - 1 ) # pivot is in elements smaller than k else: return kth_number(A__ , A__ ) if __name__ == "__main__": import doctest doctest.testmod()
16
"""simple docstring""" from __future__ import annotations def lowercase ( UpperCamelCase : list[float] ): """simple docstring""" if len(UpperCamelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) A__ : Union[str, Any] =nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
656
0
import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def lowerCamelCase( a__): _SCREAMING_SNAKE_CASE =VideoMAEConfig() set_architecture_configs(a__ ,a__) if "finetuned" not in model_name: _SCREAMING_SNAKE_CASE =False if "finetuned" in model_name: _SCREAMING_SNAKE_CASE ="huggingface/label-files" if "kinetics" in model_name: _SCREAMING_SNAKE_CASE =400 _SCREAMING_SNAKE_CASE ="kinetics400-id2label.json" elif "ssv2" in model_name: _SCREAMING_SNAKE_CASE =174 _SCREAMING_SNAKE_CASE ="something-something-v2-id2label.json" else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''') _SCREAMING_SNAKE_CASE =json.load(open(hf_hub_download(a__ ,a__ ,repo_type='''dataset''') ,'''r''')) _SCREAMING_SNAKE_CASE ={int(a__): v for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE =idalabel _SCREAMING_SNAKE_CASE ={v: k for k, v in idalabel.items()} return config def lowerCamelCase( a__ ,a__): if "small" in model_name: _SCREAMING_SNAKE_CASE =384 _SCREAMING_SNAKE_CASE =1536 _SCREAMING_SNAKE_CASE =12 _SCREAMING_SNAKE_CASE =16 _SCREAMING_SNAKE_CASE =12 _SCREAMING_SNAKE_CASE =3 _SCREAMING_SNAKE_CASE =192 _SCREAMING_SNAKE_CASE =768 elif "large" in model_name: _SCREAMING_SNAKE_CASE =1024 _SCREAMING_SNAKE_CASE =4096 _SCREAMING_SNAKE_CASE =24 _SCREAMING_SNAKE_CASE =16 _SCREAMING_SNAKE_CASE =12 _SCREAMING_SNAKE_CASE =8 _SCREAMING_SNAKE_CASE =512 _SCREAMING_SNAKE_CASE =2048 elif "huge" in model_name: _SCREAMING_SNAKE_CASE =1280 _SCREAMING_SNAKE_CASE =5120 _SCREAMING_SNAKE_CASE =32 _SCREAMING_SNAKE_CASE =16 _SCREAMING_SNAKE_CASE =12 _SCREAMING_SNAKE_CASE =8 _SCREAMING_SNAKE_CASE =640 _SCREAMING_SNAKE_CASE =2560 elif "base" not in model_name: raise ValueError('''Model name should include either \"small\", \"base\", \"large\", or \"huge\"''') def lowerCamelCase( a__): if "encoder." in name: _SCREAMING_SNAKE_CASE =name.replace('''encoder.''' ,'''''') if "cls_token" in name: _SCREAMING_SNAKE_CASE =name.replace('''cls_token''' ,'''videomae.embeddings.cls_token''') if "decoder_pos_embed" in name: _SCREAMING_SNAKE_CASE =name.replace('''decoder_pos_embed''' ,'''decoder.decoder_pos_embed''') if "pos_embed" in name and "decoder" not in name: _SCREAMING_SNAKE_CASE =name.replace('''pos_embed''' ,'''videomae.embeddings.position_embeddings''') if "patch_embed.proj" in name: _SCREAMING_SNAKE_CASE =name.replace('''patch_embed.proj''' ,'''videomae.embeddings.patch_embeddings.projection''') if "patch_embed.norm" in name: _SCREAMING_SNAKE_CASE =name.replace('''patch_embed.norm''' ,'''videomae.embeddings.norm''') if "decoder.blocks" in name: _SCREAMING_SNAKE_CASE =name.replace('''decoder.blocks''' ,'''decoder.decoder_layers''') if "blocks" in name: _SCREAMING_SNAKE_CASE =name.replace('''blocks''' ,'''videomae.encoder.layer''') if "attn.proj" in name: _SCREAMING_SNAKE_CASE =name.replace('''attn.proj''' ,'''attention.output.dense''') if "attn" in name and "bias" not in name: _SCREAMING_SNAKE_CASE =name.replace('''attn''' ,'''attention.self''') if "attn" in name: _SCREAMING_SNAKE_CASE =name.replace('''attn''' ,'''attention.attention''') if "norm1" in name: _SCREAMING_SNAKE_CASE =name.replace('''norm1''' ,'''layernorm_before''') if "norm2" in name: _SCREAMING_SNAKE_CASE =name.replace('''norm2''' ,'''layernorm_after''') if "mlp.fc1" in name: _SCREAMING_SNAKE_CASE =name.replace('''mlp.fc1''' ,'''intermediate.dense''') if "mlp.fc2" in name: _SCREAMING_SNAKE_CASE =name.replace('''mlp.fc2''' ,'''output.dense''') if "decoder_embed" in name: _SCREAMING_SNAKE_CASE =name.replace('''decoder_embed''' ,'''decoder.decoder_embed''') if "decoder_norm" in name: _SCREAMING_SNAKE_CASE =name.replace('''decoder_norm''' ,'''decoder.decoder_norm''') if "decoder_pred" in name: _SCREAMING_SNAKE_CASE =name.replace('''decoder_pred''' ,'''decoder.decoder_pred''') if "norm.weight" in name and "decoder" not in name and "fc" not in name: _SCREAMING_SNAKE_CASE =name.replace('''norm.weight''' ,'''videomae.layernorm.weight''') if "norm.bias" in name and "decoder" not in name and "fc" not in name: _SCREAMING_SNAKE_CASE =name.replace('''norm.bias''' ,'''videomae.layernorm.bias''') if "head" in name and "decoder" not in name: _SCREAMING_SNAKE_CASE =name.replace('''head''' ,'''classifier''') return name def lowerCamelCase( a__ ,a__): for key in orig_state_dict.copy().keys(): _SCREAMING_SNAKE_CASE =orig_state_dict.pop(a__) if key.startswith('''encoder.'''): _SCREAMING_SNAKE_CASE =key.replace('''encoder.''' ,'''''') if "qkv" in key: _SCREAMING_SNAKE_CASE =key.split('''.''') if key.startswith('''decoder.blocks'''): _SCREAMING_SNAKE_CASE =config.decoder_hidden_size _SCREAMING_SNAKE_CASE =int(key_split[2]) _SCREAMING_SNAKE_CASE ="decoder.decoder_layers." if "weight" in key: _SCREAMING_SNAKE_CASE =val[:dim, :] _SCREAMING_SNAKE_CASE =val[dim : dim * 2, :] _SCREAMING_SNAKE_CASE =val[-dim:, :] else: _SCREAMING_SNAKE_CASE =config.hidden_size _SCREAMING_SNAKE_CASE =int(key_split[1]) _SCREAMING_SNAKE_CASE ="videomae.encoder.layer." if "weight" in key: _SCREAMING_SNAKE_CASE =val[:dim, :] _SCREAMING_SNAKE_CASE =val[dim : dim * 2, :] _SCREAMING_SNAKE_CASE =val[-dim:, :] else: _SCREAMING_SNAKE_CASE =val return orig_state_dict def lowerCamelCase( ): _SCREAMING_SNAKE_CASE =hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' ,filename='''eating_spaghetti.npy''' ,repo_type='''dataset''') _SCREAMING_SNAKE_CASE =np.load(a__) return list(a__) def lowerCamelCase( a__ ,a__ ,a__ ,a__): _SCREAMING_SNAKE_CASE =get_videomae_config(a__) if "finetuned" in model_name: _SCREAMING_SNAKE_CASE =VideoMAEForVideoClassification(a__) else: _SCREAMING_SNAKE_CASE =VideoMAEForPreTraining(a__) # download original checkpoint, hosted on Google Drive _SCREAMING_SNAKE_CASE ="pytorch_model.bin" gdown.cached_download(a__ ,a__ ,quiet=a__) _SCREAMING_SNAKE_CASE =torch.load(a__ ,map_location='''cpu''') if "model" in files: _SCREAMING_SNAKE_CASE =files["model"] else: _SCREAMING_SNAKE_CASE =files["module"] _SCREAMING_SNAKE_CASE =convert_state_dict(a__ ,a__) model.load_state_dict(a__) model.eval() # verify model on basic input _SCREAMING_SNAKE_CASE =VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] ,image_std=[0.5, 0.5, 0.5]) _SCREAMING_SNAKE_CASE =prepare_video() _SCREAMING_SNAKE_CASE =image_processor(a__ ,return_tensors='''pt''') if "finetuned" not in model_name: _SCREAMING_SNAKE_CASE =hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' ,filename='''bool_masked_pos.pt''') _SCREAMING_SNAKE_CASE =torch.load(a__) _SCREAMING_SNAKE_CASE =model(**a__) _SCREAMING_SNAKE_CASE =outputs.logits _SCREAMING_SNAKE_CASE =[ "videomae-small-finetuned-kinetics", "videomae-small-finetuned-ssv2", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) "videomae-base-short", "videomae-base-short-finetuned-kinetics", "videomae-base", "videomae-base-finetuned-kinetics", "videomae-large", "videomae-large-finetuned-kinetics", "videomae-huge-finetuned-kinetics", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) "videomae-base-short-ssv2", "videomae-base-short-finetuned-ssv2", "videomae-base-ssv2", "videomae-base-finetuned-ssv2", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": _SCREAMING_SNAKE_CASE =torch.Size([1, 400]) _SCREAMING_SNAKE_CASE =torch.tensor([-0.9291, -0.4061, -0.9307]) elif model_name == "videomae-small-finetuned-ssv2": _SCREAMING_SNAKE_CASE =torch.Size([1, 174]) _SCREAMING_SNAKE_CASE =torch.tensor([0.2671, -0.4689, -0.8235]) elif model_name == "videomae-base": _SCREAMING_SNAKE_CASE =torch.Size([1, 1408, 1536]) _SCREAMING_SNAKE_CASE =torch.tensor([[0.7739, 0.7968, 0.7089], [0.6701, 0.7487, 0.6209], [0.4287, 0.5158, 0.4773]]) elif model_name == "videomae-base-short": _SCREAMING_SNAKE_CASE =torch.Size([1, 1408, 1536]) _SCREAMING_SNAKE_CASE =torch.tensor([[0.7994, 0.9612, 0.8508], [0.7401, 0.8958, 0.8302], [0.5862, 0.7468, 0.7325]]) # we verified the loss both for normalized and unnormalized targets for this one _SCREAMING_SNAKE_CASE =torch.tensor([0.5142]) if config.norm_pix_loss else torch.tensor([0.6469]) elif model_name == "videomae-large": _SCREAMING_SNAKE_CASE =torch.Size([1, 1408, 1536]) _SCREAMING_SNAKE_CASE =torch.tensor([[0.7149, 0.7997, 0.6966], [0.6768, 0.7869, 0.6948], [0.5139, 0.6221, 0.5605]]) elif model_name == "videomae-large-finetuned-kinetics": _SCREAMING_SNAKE_CASE =torch.Size([1, 400]) _SCREAMING_SNAKE_CASE =torch.tensor([0.0771, 0.0011, -0.3625]) elif model_name == "videomae-huge-finetuned-kinetics": _SCREAMING_SNAKE_CASE =torch.Size([1, 400]) _SCREAMING_SNAKE_CASE =torch.tensor([0.2433, 0.1632, -0.4894]) elif model_name == "videomae-base-short-finetuned-kinetics": _SCREAMING_SNAKE_CASE =torch.Size([1, 400]) _SCREAMING_SNAKE_CASE =torch.tensor([0.6588, 0.0990, -0.2493]) elif model_name == "videomae-base-finetuned-kinetics": _SCREAMING_SNAKE_CASE =torch.Size([1, 400]) _SCREAMING_SNAKE_CASE =torch.tensor([0.3669, -0.0688, -0.2421]) elif model_name == "videomae-base-short-ssv2": _SCREAMING_SNAKE_CASE =torch.Size([1, 1408, 1536]) _SCREAMING_SNAKE_CASE =torch.tensor([[0.4712, 0.5296, 0.5786], [0.2278, 0.2729, 0.4026], [0.0352, 0.0730, 0.2506]]) elif model_name == "videomae-base-short-finetuned-ssv2": _SCREAMING_SNAKE_CASE =torch.Size([1, 174]) _SCREAMING_SNAKE_CASE =torch.tensor([-0.0537, -0.1539, -0.3266]) elif model_name == "videomae-base-ssv2": _SCREAMING_SNAKE_CASE =torch.Size([1, 1408, 1536]) _SCREAMING_SNAKE_CASE =torch.tensor([[0.8131, 0.8727, 0.8546], [0.7366, 0.9377, 0.8870], [0.5935, 0.8874, 0.8564]]) elif model_name == "videomae-base-finetuned-ssv2": _SCREAMING_SNAKE_CASE =torch.Size([1, 174]) _SCREAMING_SNAKE_CASE =torch.tensor([0.1961, -0.8337, -0.6389]) else: raise ValueError(f"Model name not supported. Should be one of {model_names}") # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] ,a__ ,atol=1e-4) else: print('''Logits:''' ,logits[0, :3, :3]) assert torch.allclose(logits[0, :3, :3] ,a__ ,atol=1e-4) print('''Logits ok!''') # verify loss, if applicable if model_name == "videomae-base-short": _SCREAMING_SNAKE_CASE =outputs.loss assert torch.allclose(a__ ,a__ ,atol=1e-4) print('''Loss ok!''') if pytorch_dump_folder_path is not None: print(f"Saving model and image processor to {pytorch_dump_folder_path}") image_processor.save_pretrained(a__) model.save_pretrained(a__) if push_to_hub: print('''Pushing to the hub...''') model.push_to_hub(a__ ,organization='''nielsr''') if __name__ == "__main__": snake_case_ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4''', type=str, help=( '''URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct''' ''' download link.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''/Users/nielsrogge/Documents/VideoMAE/Test''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--model_name''', default='''videomae-base''', type=str, help='''Name of the model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) snake_case_ : Optional[Any] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
691
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : Optional[Any] = { "configuration_mega": ["MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegaConfig", "MegaOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ "MEGA_PRETRAINED_MODEL_ARCHIVE_LIST", "MegaForCausalLM", "MegaForMaskedLM", "MegaForMultipleChoice", "MegaForQuestionAnswering", "MegaForSequenceClassification", "MegaForTokenClassification", "MegaModel", "MegaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
from functools import reduce _lowerCamelCase : Tuple = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def a_ ( __lowercase : str = N ) -> Optional[int]: return max( # mypy cannot properly interpret reduce int(reduce(lambda __lowercase , __lowercase : str(int(__lowercase ) * int(__lowercase ) ) , n[i : i + 13] ) ) for i in range(len(__lowercase ) - 12 ) ) if __name__ == "__main__": print(F'{solution() = }')
686
"""simple docstring""" def lowercase ( UpperCamelCase : int ): """simple docstring""" if num <= 0: raise ValueError("Input must be a positive integer" ) A__ : Union[str, Any] =[True] * (num + 1) A__ : Union[str, Any] =2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCamelCase ): A__ : str =False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
656
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def snake_case ( self ): __lowerCAmelCase = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) __lowerCAmelCase = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house __lowerCAmelCase = torch.Size((1, 12, 7_68) ) # batch_size, sequence_length, embedding_vector_dim __lowerCAmelCase = torch.tensor( [[-0.0_1_0_1, 0.1_2_1_8, -0.0_8_0_3, 0.0_8_0_1, 0.1_3_2_7, 0.0_7_7_6, -0.1_2_1_5, 0.2_3_8_3, 0.3_3_3_8, 0.3_1_0_6, 0.0_3_0_0, 0.0_2_5_2]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __lowerCAmelCase = model(UpperCamelCase__ )["last_hidden_state"].detach() self.assertEqual(output.shape , UpperCamelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , UpperCamelCase__ , atol=1e-3 ) ) @slow def snake_case ( self ): __lowerCAmelCase = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) __lowerCAmelCase = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house __lowerCAmelCase = torch.Size((1, 12, 10_24) ) # batch_size, sequence_length, embedding_vector_dim __lowerCAmelCase = torch.tensor( [[-0.0_6_9_9, -0.0_3_1_8, 0.0_7_0_5, -0.1_2_4_1, 0.0_9_9_9, -0.0_5_2_0, 0.1_0_0_4, -0.1_8_3_8, -0.4_7_0_4, 0.1_4_3_7, 0.0_8_2_1, 0.0_1_2_6]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __lowerCAmelCase = model(UpperCamelCase__ )["last_hidden_state"].detach() self.assertEqual(output.shape , UpperCamelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , UpperCamelCase__ , atol=1e-3 ) )
636
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : List[Any] ): A__ : Tuple =torch.nn.Linear(10 , 10 ) A__ : List[str] =torch.optim.SGD(model.parameters() , 0.1 ) A__ : Union[str, Any] =Accelerator() A__ : str =accelerator.prepare(UpperCamelCase__ ) try: pickle.loads(pickle.dumps(UpperCamelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
"""simple docstring""" def _snake_case ( snake_case__ : str ): assert column_title.isupper() A = 0 A = len(snake_case__ ) - 1 A = 0 while index >= 0: A = (ord(column_title[index] ) - 64) * pow(26 , snake_case__ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
91
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __A : Optional[int] = None __A : Union[str, Any] = logging.get_logger(__name__) __A : List[Any] = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __A : str = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } __A : List[str] = { "google/bigbird-roberta-base": 4_096, "google/bigbird-roberta-large": 4_096, "google/bigbird-base-trivia-itc": 4_096, } __A : Tuple = "▁" class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Dict = VOCAB_FILES_NAMES __magic_name__ : Any = PRETRAINED_VOCAB_FILES_MAP __magic_name__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ : List[Any] = BigBirdTokenizer __magic_name__ : Any = ["""input_ids""", """attention_mask"""] __magic_name__ : List[int] = [] def __init__( self : str , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : int="</s>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[Any]="[SEP]" , UpperCamelCase__ : List[Any]="[MASK]" , UpperCamelCase__ : str="[CLS]" , **UpperCamelCase__ : List[Any] , ): A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else bos_token A__ : Optional[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else eos_token A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else unk_token A__ : int =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else pad_token A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else cls_token A__ : List[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) A__ : List[Any] =vocab_file A__ : Optional[int] =False if not self.vocab_file else True def _UpperCAmelCase ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1] def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : Dict =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : List[str] =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
656
0
"""simple docstring""" def _UpperCamelCase ( A , A , A ): if len(A ) != len(A ): raise ValueError("The length of profit and weight must be same." ) if max_weight <= 0: raise ValueError("max_weight must greater than zero." ) if any(p < 0 for p in profit ): raise ValueError("Profit can not be negative." ) if any(w < 0 for w in weight ): raise ValueError("Weight can not be negative." ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. UpperCamelCase_ =[p / w for p, w in zip(A , A )] # Creating a copy of the list and sorting profit/weight in ascending order UpperCamelCase_ =sorted(A ) # declaring useful variables UpperCamelCase_ =len(A ) UpperCamelCase_ =0 UpperCamelCase_ =0 UpperCamelCase_ =0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight UpperCamelCase_ =sorted_profit_by_weight[length - i - 1] UpperCamelCase_ =profit_by_weight.index(A ) UpperCamelCase_ =-1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( "Input profits, weights, and then max_weight (all positive ints) separated by " "spaces." ) A_ = [int(x) for x in input("Input profits separated by spaces: ").split()] A_ = [int(x) for x in input("Input weights separated by spaces: ").split()] A_ = int(input("Max weight allowed: ")) # Function Call calc_profit(profit, weight, max_weight)
391
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Optional[int] = logging.get_logger(__name__) __A : Optional[int] = {"vocab_file": "spiece.model"} __A : List[Any] = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : Dict="<s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : Optional[int]="<sep>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[int]="<cls>" , UpperCamelCase__ : List[str]="<mask>" , UpperCamelCase__ : Optional[Any]=["<eop>", "<eod>"] , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : Dict , ): A__ : List[str] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token A__ : Tuple ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) A__ : Dict =3 A__ : int =do_lower_case A__ : str =remove_space A__ : Optional[Any] =keep_accents A__ : int =vocab_file A__ : Dict =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) A__ : Union[str, Any] =jieba A__ : List[str] =str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _UpperCAmelCase ( self : Union[str, Any] ): return len(self.sp_model ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Any ={self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[str] ): A__ : Union[str, Any] =self.__dict__.copy() A__ : Tuple =None return state def __setstate__( self : Tuple , UpperCamelCase__ : int ): A__ : Union[str, Any] =d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A__ : Optional[int] ={} A__ : Union[str, Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Dict ): if self.remove_space: A__ : Optional[int] =" ".join(inputs.strip().split() ) else: A__ : Optional[Any] =inputs A__ : Any =outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: A__ : Optional[Any] =unicodedata.normalize("NFKD" , UpperCamelCase__ ) A__ : Tuple ="".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: A__ : str =outputs.lower() return outputs def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : str ): A__ : Optional[int] =self.preprocess_text(UpperCamelCase__ ) A__ : Dict =self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) A__ : List[str] =[] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): A__ : str =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : Union[str, Any] =cur_pieces[1:] else: A__ : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def _UpperCAmelCase ( self : int , UpperCamelCase__ : str ): return self.sp_model.PieceToId(UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[Any] ): return self.sp_model.IdToPiece(UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : str ): A__ : Optional[int] ="".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def _UpperCAmelCase ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : Optional[Any] =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : Tuple =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: A__ : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def _UpperCAmelCase ( self : str , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : int ): A__ : List[Any] =super()._decode(*UpperCamelCase__ , **UpperCamelCase__ ) A__ : Union[str, Any] =text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
656
0
'''simple docstring''' from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def a_ ( ): A_ = HfArgumentParser(UpperCamelCase_ ) A_ = parser.parse_args_into_dataclasses()[0] A_ = TensorFlowBenchmark(args=UpperCamelCase_ ) try: A_ = parser.parse_args_into_dataclasses()[0] except ValueError as e: A_ = "Arg --no_{0} is no longer used, please use --no-{0} instead." A_ = " ".join(str(UpperCamelCase_ ).split(" " )[:-1] ) A_ = "" A_ = eval(str(UpperCamelCase_ ).split(" " )[-1] ) A_ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCamelCase_ ) if len(UpperCamelCase_ ) > 0: A_ = full_error_msg + begin_error_msg + str(UpperCamelCase_ ) raise ValueError(UpperCamelCase_ ) benchmark.run() if __name__ == "__main__": main()
452
"""simple docstring""" def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations(UpperCamelCase : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations_with_dp_array( UpperCamelCase : int , UpperCamelCase : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] A__ : str =sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase ) for item in array ) A__ : List[str] =answer return answer A__ : List[Any] =[-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase , UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" A__ : str =[0] * (target + 1) A__ : Optional[Any] =1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[Any] = 3 __A : Optional[Any] = 5 __A : int = [1, 2, 5] print(combination_sum_iv(n, array, target))
656
0
"""simple docstring""" class __snake_case : def __init__( self: Optional[int] ): __lowerCamelCase = "" __lowerCamelCase = "" __lowerCamelCase = [] def __a ( self: Tuple , A_: int , A_: int ): if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: __lowerCamelCase = self.__min_dist_top_down_dp(UpperCamelCase__ , n - 1 ) __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , UpperCamelCase__ ) __lowerCamelCase = self.__min_dist_top_down_dp(m - 1 , n - 1 ) __lowerCamelCase = 1 + min(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return self.dp[m][n] def __a ( self: Dict , A_: str , A_: str ): __lowerCamelCase = worda __lowerCamelCase = worda __lowerCamelCase = [[-1 for _ in range(len(UpperCamelCase__ ) )] for _ in range(len(UpperCamelCase__ ) )] return self.__min_dist_top_down_dp(len(UpperCamelCase__ ) - 1 , len(UpperCamelCase__ ) - 1 ) def __a ( self: Optional[int] , A_: str , A_: str ): __lowerCamelCase = worda __lowerCamelCase = worda __lowerCamelCase = len(UpperCamelCase__ ) __lowerCamelCase = len(UpperCamelCase__ ) __lowerCamelCase = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty __lowerCamelCase = j elif j == 0: # second string is empty __lowerCamelCase = i elif worda[i - 1] == worda[j - 1]: # last characters are equal __lowerCamelCase = self.dp[i - 1][j - 1] else: __lowerCamelCase = self.dp[i][j - 1] __lowerCamelCase = self.dp[i - 1][j] __lowerCamelCase = self.dp[i - 1][j - 1] __lowerCamelCase = 1 + min(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return self.dp[m][n] if __name__ == "__main__": __magic_name__ : str = EditDistance() print('****************** Testing Edit Distance DP Algorithm ******************') print() __magic_name__ : str = input('Enter the first string: ').strip() __magic_name__ : List[Any] = input('Enter the second string: ').strip() print() print(f"""The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}""") print(f"""The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}""") print() print('*************** End of Testing Edit Distance DP Algorithm ***************')
281
"""simple docstring""" import math import tensorflow as tf from packaging import version def lowercase ( UpperCamelCase : Optional[Any] ): """simple docstring""" A__ : List[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : Optional[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : Tuple =tf.cast(math.pi , x.dtype ) A__ : Dict =tf.cast(0.04_47_15 , x.dtype ) A__ : Optional[int] =0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(UpperCamelCase , 3 )) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) return x * tf.tanh(tf.math.softplus(UpperCamelCase ) ) def lowercase ( UpperCamelCase : List[str] ): """simple docstring""" A__ : Union[str, Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =tf.cast(0.04_47_15 , x.dtype ) A__ : List[Any] =tf.cast(0.79_78_84_56_08 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) A__ : str =tf.cast(1.7_02 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" return tf.clip_by_value(_gelu(UpperCamelCase ) , -10 , 10 ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any=-1 ): """simple docstring""" A__ , A__ : Optional[Any] =tf.split(UpperCamelCase , 2 , axis=UpperCamelCase ) return a * tf.math.sigmoid(UpperCamelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def lowercase ( UpperCamelCase : int ): """simple docstring""" return tf.keras.activations.gelu(UpperCamelCase , approximate=UpperCamelCase ) __A : Optional[Any] = tf.keras.activations.gelu __A : Optional[Any] = approximate_gelu_wrap else: __A : Any = _gelu __A : Union[str, Any] = _gelu_new __A : List[str] = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F'''function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}''' )
656
0
from scipy.stats import spearmanr import datasets _UpperCamelCase : List[str] ="\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n" _UpperCamelCase : Tuple ="\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {'spearmanr': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results['spearmanr'])\n -0.7\n >>> print(round(results['spearmanr_pvalue'], 2))\n 0.19\n" _UpperCamelCase : Union[str, Any] =R"\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase__ ( datasets.Metric ): def A__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def A__ ( self ,A__ ,A__ ,A__=False ): _A : Optional[Any] = spearmanr(UpperCamelCase__ ,UpperCamelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
206
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict ): A__ : Optional[Any] =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_attention_heads" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_encoder_blocks" ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any]=13 , UpperCamelCase__ : Tuple=64 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Union[str, Any]=4 , UpperCamelCase__ : Dict=[2, 2, 2, 2] , UpperCamelCase__ : Union[str, Any]=[8, 4, 2, 1] , UpperCamelCase__ : Tuple=[16, 32, 64, 128] , UpperCamelCase__ : Optional[int]=[1, 4, 8, 16] , UpperCamelCase__ : Any=[1, 2, 4, 8] , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Dict="gelu" , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[Any]=None , ): A__ : Tuple =parent A__ : List[Any] =batch_size A__ : List[Any] =image_size A__ : Union[str, Any] =num_channels A__ : Optional[int] =num_encoder_blocks A__ : Any =sr_ratios A__ : Any =depths A__ : List[Any] =hidden_sizes A__ : List[Any] =downsampling_rates A__ : List[str] =num_attention_heads A__ : int =is_training A__ : List[Any] =use_labels A__ : Any =hidden_act A__ : Dict =hidden_dropout_prob A__ : int =attention_probs_dropout_prob A__ : List[Any] =initializer_range A__ : Tuple =num_labels A__ : List[Any] =scope def _UpperCAmelCase ( self : Optional[int] ): A__ : List[str] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Any =None if self.use_labels: A__ : Tuple =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A__ : List[Any] =self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self : Tuple ): return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int ): A__ : Any =SegformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Dict =model(UpperCamelCase__ ) A__ : Optional[int] =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[Any] ): A__ : str =self.num_labels A__ : Optional[Any] =SegformerForSemanticSegmentation(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Optional[Any] =model(UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) A__ : List[Any] =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str ): A__ : Tuple =1 A__ : Tuple =SegformerForSemanticSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : List[str] =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(UpperCamelCase__ ) A__ : Dict =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : str ): A__ : Union[str, Any] =self.prepare_config_and_inputs() A__ , A__ , A__ : Tuple =config_and_inputs A__ : Tuple ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Dict = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __magic_name__ : Optional[int] = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ : Dict = True __magic_name__ : List[str] = False __magic_name__ : Optional[Any] = False __magic_name__ : str = False def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =SegformerModelTester(self ) A__ : Tuple =SegformerConfigTester(self , config_class=UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): self.config_tester.run_common_tests() def _UpperCAmelCase ( self : Dict ): A__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*UpperCamelCase__ ) @unittest.skip("SegFormer does not use inputs_embeds" ) def _UpperCAmelCase ( self : Dict ): pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods" ) def _UpperCAmelCase ( self : Tuple ): pass def _UpperCAmelCase ( self : List[str] ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : int =model_class(UpperCamelCase__ ) A__ : Optional[int] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Optional[int] =[*signature.parameters.keys()] A__ : List[str] =["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() A__ : Union[str, Any] =True for model_class in self.all_model_classes: A__ : Optional[Any] =True A__ : Union[str, Any] =False A__ : str =True A__ : Optional[int] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : str =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Any =outputs.attentions A__ : List[str] =sum(self.model_tester.depths ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ : Dict =True A__ : str =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Any =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Union[str, Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : List[Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) A__ : Tuple =(self.model_tester.image_size // 32) ** 2 A__ : Optional[Any] =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) A__ : int =len(UpperCamelCase__ ) # Check attention is always last and order is fine A__ : Optional[Any] =True A__ : Any =True A__ : Union[str, Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Optional[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(out_len + 1 , len(UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : Union[str, Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _UpperCAmelCase ( self : List[Any] ): def check_hidden_states_output(UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple ): A__ : Optional[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : List[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.hidden_states A__ : int =self.model_tester.num_encoder_blocks self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Optional[Any] =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : str =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[int] ): if not self.model_tester.is_training: return A__ , A__ : int =self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] =True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ): continue A__ : List[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() A__ : int =self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) A__ : Union[str, Any] =model(**UpperCamelCase__ ).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _UpperCAmelCase ( self : Tuple ): pass @slow def _UpperCAmelCase ( self : Tuple ): for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Tuple =SegformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowercase ( ): """simple docstring""" A__ : List[Any] =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): # only resize + normalize A__ : List[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : Union[str, Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : Union[str, Any] =prepare_img() A__ : Union[str, Any] =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : int =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : Dict =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : Optional[int] =torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : Union[str, Any] ): # only resize + normalize A__ : Dict =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : int =SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024" ).to(UpperCamelCase__ ) A__ : Tuple =prepare_img() A__ : str =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Optional[int] =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : List[str] =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : List[Any] =torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-1 ) ) @slow def _UpperCAmelCase ( self : int ): # only resize + normalize A__ : Optional[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : List[Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : str =prepare_img() A__ : Dict =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Any =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : Dict =model(UpperCamelCase__ ) A__ : Any =outputs.logits.detach().cpu() A__ : Union[str, Any] =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ , target_sizes=[(500, 300)] ) A__ : List[str] =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ ) A__ : int =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ ) A__ : Tuple =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ )
656
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" __snake_case = args.pruning_method __snake_case = args.threshold __snake_case = args.model_name_or_path.rstrip("/" ) __snake_case = args.target_model_path print(F'''Load fine-pruned model from {model_name_or_path}''' ) __snake_case = torch.load(os.path.join(SCREAMING_SNAKE_CASE , "pytorch_model.bin" ) ) __snake_case = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: __snake_case = tensor print(F'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: __snake_case = tensor print(F'''Copied layer {name}''' ) elif "bias" in name: __snake_case = tensor print(F'''Copied layer {name}''' ) else: if pruning_method == "magnitude": __snake_case = MagnitudeBinarizer.apply(inputs=SCREAMING_SNAKE_CASE , threshold=SCREAMING_SNAKE_CASE ) __snake_case = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue __snake_case = name[:-6] __snake_case = model[F'''{prefix_}mask_scores'''] __snake_case = TopKBinarizer.apply(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __snake_case = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue __snake_case = name[:-6] __snake_case = model[F'''{prefix_}mask_scores'''] __snake_case = ThresholdBinarizer.apply(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __snake_case = tensor * mask print(F'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue __snake_case = name[:-6] __snake_case = model[F'''{prefix_}mask_scores'''] __snake_case = -0.1, 1.1 __snake_case = torch.sigmoid(SCREAMING_SNAKE_CASE ) __snake_case = s * (r - l) + l __snake_case = s_bar.clamp(min=0.0 , max=1.0 ) __snake_case = tensor * mask print(F'''Pruned layer {name}''' ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: __snake_case = os.path.join( os.path.dirname(SCREAMING_SNAKE_CASE ) , F'''bertarized_{os.path.basename(SCREAMING_SNAKE_CASE )}''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): shutil.copytree(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print(F'''\nCreated folder {target_model_path}''' ) torch.save(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) _SCREAMING_SNAKE_CASE = parser.parse_args() main(args)
163
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any]=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[str]=99 , UpperCamelCase__ : Optional[Any]=32 , UpperCamelCase__ : Any=5 , UpperCamelCase__ : Dict=4 , UpperCamelCase__ : Union[str, Any]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Any=0.1 , UpperCamelCase__ : Optional[Any]=0.1 , UpperCamelCase__ : Optional[Any]=512 , UpperCamelCase__ : Union[str, Any]=16 , UpperCamelCase__ : List[str]=2 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : List[Any]=4 , ): A__ : str =parent A__ : List[str] =batch_size A__ : Any =seq_length A__ : List[str] =is_training A__ : List[Any] =use_attention_mask A__ : List[Any] =use_token_type_ids A__ : Dict =use_labels A__ : List[Any] =vocab_size A__ : Optional[int] =hidden_size A__ : Optional[Any] =num_hidden_layers A__ : str =num_attention_heads A__ : int =intermediate_size A__ : Tuple =hidden_act A__ : Tuple =hidden_dropout_prob A__ : Dict =attention_probs_dropout_prob A__ : Any =max_position_embeddings A__ : Any =type_vocab_size A__ : Union[str, Any] =type_sequence_label_size A__ : Optional[Any] =initializer_range A__ : int =num_choices def _UpperCAmelCase ( self : Tuple ): A__ : List[str] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[str] =None if self.use_attention_mask: A__ : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) A__ : str =None if self.use_token_type_ids: A__ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Any =RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCAmelCase ( self : Tuple ): A__ : Dict =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : str =config_and_inputs A__ : Optional[Any] ={"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _UpperCAmelCase ( self : int ): A__ : str =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Union[str, Any] =config_and_inputs A__ : Union[str, Any] =True A__ : List[Any] =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : Tuple =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __lowerCAmelCase ( _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Union[str, Any] = True __magic_name__ : Dict = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Optional[int] =FlaxRobertaPreLayerNormModelTester(self ) @slow def _UpperCAmelCase ( self : List[Any] ): for model_class_name in self.all_model_classes: A__ : Tuple =model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Union[str, Any] =model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): A__ : Any =FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Tuple =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : str =model(UpperCamelCase__ )[0] A__ : List[Any] =[1, 11, 50265] self.assertEqual(list(output.shape ) , UpperCamelCase__ ) # compare the actual values for a slice. A__ : Any =np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : List[Any] ): A__ : Union[str, Any] =FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : List[Any] =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : Dict =model(UpperCamelCase__ )[0] # compare the actual values for a slice. A__ : Optional[Any] =np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) )
656
0
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class _UpperCAmelCase : """simple docstring""" def __init__( self , _lowercase , _lowercase=99 , _lowercase=13 , _lowercase=7 , _lowercase=9 , _lowercase=True , _lowercase=True , _lowercase=False , _lowercase=32 , _lowercase=5 , _lowercase=4 , _lowercase=37 , _lowercase=8 , _lowercase=0.1 , _lowercase=0.002 , _lowercase=1 , _lowercase=0 , _lowercase=0 , _lowercase=None , _lowercase=None , ) -> Dict: _lowerCamelCase : Any = parent _lowerCamelCase : Optional[Any] = batch_size _lowerCamelCase : int = encoder_seq_length _lowerCamelCase : Optional[Any] = decoder_seq_length # For common tests _lowerCamelCase : str = self.decoder_seq_length _lowerCamelCase : Any = is_training _lowerCamelCase : Union[str, Any] = use_attention_mask _lowerCamelCase : int = use_labels _lowerCamelCase : str = vocab_size _lowerCamelCase : Optional[Any] = hidden_size _lowerCamelCase : Any = num_hidden_layers _lowerCamelCase : Optional[int] = num_attention_heads _lowerCamelCase : Tuple = d_ff _lowerCamelCase : Dict = relative_attention_num_buckets _lowerCamelCase : Any = dropout_rate _lowerCamelCase : List[Any] = initializer_factor _lowerCamelCase : Tuple = eos_token_id _lowerCamelCase : Union[str, Any] = pad_token_id _lowerCamelCase : List[str] = decoder_start_token_id _lowerCamelCase : str = None _lowerCamelCase : List[str] = decoder_layers def a__ ( self ) -> str: return TaConfig.from_pretrained('''google/umt5-base''' ) def a__ ( self , _lowercase , _lowercase , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> int: if attention_mask is None: _lowerCamelCase : List[str] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: _lowerCamelCase : Union[str, Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: _lowerCamelCase : Any = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCamelCase__ ) if decoder_head_mask is None: _lowerCamelCase : Union[str, Any] = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCamelCase__ ) if cross_attn_head_mask is None: _lowerCamelCase : Optional[int] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCamelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ) -> Dict: _lowerCamelCase : str = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) _lowerCamelCase : Optional[int] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input _lowerCamelCase : Tuple = input_ids.clamp(self.pad_token_id + 1 ) _lowerCamelCase : Optional[Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) _lowerCamelCase : str = self.get_config() _lowerCamelCase : List[Any] = config.num_attention_heads _lowerCamelCase : List[str] = self.prepare_inputs_dict(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return config, input_dict def a__ ( self ) -> Optional[Any]: _lowerCamelCase : Dict = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ) -> Tuple: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ) -> Tuple: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , ) -> Optional[int]: _lowerCamelCase : Optional[Any] = UMTaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _lowerCamelCase : List[str] = model( input_ids=UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ , attention_mask=UpperCamelCase__ , decoder_attention_mask=UpperCamelCase__ , ) _lowerCamelCase : Dict = model(input_ids=UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ ) _lowerCamelCase : Optional[Any] = result.last_hidden_state _lowerCamelCase : Union[str, Any] = result.past_key_values _lowerCamelCase : Optional[int] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCamelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , ) -> Any: _lowerCamelCase : Any = UMTaModel(config=UpperCamelCase__ ).get_decoder().to(UpperCamelCase__ ).eval() # first forward pass _lowerCamelCase : List[str] = model(UpperCamelCase__ , use_cache=UpperCamelCase__ ) _lowerCamelCase : str = model(UpperCamelCase__ ) _lowerCamelCase : Optional[Any] = model(UpperCamelCase__ , use_cache=UpperCamelCase__ ) self.parent.assertTrue(len(UpperCamelCase__ ) == len(UpperCamelCase__ ) ) self.parent.assertTrue(len(UpperCamelCase__ ) == len(UpperCamelCase__ ) + 1 ) _lowerCamelCase : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _lowerCamelCase : Dict = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and _lowerCamelCase : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase : Any = model(UpperCamelCase__ )["last_hidden_state"] _lowerCamelCase : Optional[int] = model(UpperCamelCase__ , past_key_values=UpperCamelCase__ )["last_hidden_state"] # select random slice _lowerCamelCase : Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() _lowerCamelCase : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) ) def a__ ( self , _lowercase , _lowercase , ) -> Optional[Any]: _lowerCamelCase : str = UMTaModel(config=UpperCamelCase__ ).to(UpperCamelCase__ ).half().eval() _lowerCamelCase : int = model(**UpperCamelCase__ )["last_hidden_state"] self.parent.assertFalse(torch.isnan(UpperCamelCase__ ).any().item() ) @require_torch class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" __snake_case = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) __snake_case = (UMTaForConditionalGeneration,) if is_torch_available() else () __snake_case = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) __snake_case = True __snake_case = False __snake_case = False __snake_case = True __snake_case = True # The small UMT5 model needs higher percentages for CPU/MP tests __snake_case = [0.8, 0.9] def a__ ( self ) -> Tuple: _lowerCamelCase : Any = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def a__ ( self ) -> Any: _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() _lowerCamelCase : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCamelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F'''{tmpdirname}/t5_test.onnx''' , export_params=UpperCamelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def a__ ( self ) -> int: _lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCamelCase__ ) def a__ ( self ) -> str: _lowerCamelCase : Optional[Any] = ["encoder_attentions", "decoder_attentions", "cross_attentions"] _lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() _lowerCamelCase : Optional[int] = config_and_inputs[0] _lowerCamelCase : List[Any] = UMTaForConditionalGeneration(UpperCamelCase__ ).eval() model.to(UpperCamelCase__ ) _lowerCamelCase : int = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=UpperCamelCase__ ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCamelCase__ ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCamelCase__ ), } for attn_name, (name, mask) in zip(UpperCamelCase__ , head_masking.items() ): _lowerCamelCase : Tuple = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": _lowerCamelCase : Dict = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCamelCase__ ) _lowerCamelCase : str = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCamelCase__ , return_dict_in_generate=UpperCamelCase__ , **UpperCamelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step _lowerCamelCase : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def a__ ( self ) -> int: pass @require_torch @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def a__ ( self ) -> Union[str, Any]: _lowerCamelCase : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCamelCase__ ).to(UpperCamelCase__ ) _lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCamelCase__ , legacy=UpperCamelCase__ ) _lowerCamelCase : int = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] _lowerCamelCase : Optional[Any] = tokenizer(UpperCamelCase__ , return_tensors='''pt''' , padding=UpperCamelCase__ ).input_ids # fmt: off _lowerCamelCase : List[str] = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCamelCase__ , UpperCamelCase__ ) _lowerCamelCase : Optional[int] = model.generate(input_ids.to(UpperCamelCase__ ) ) _lowerCamelCase : int = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] _lowerCamelCase : Optional[int] = tokenizer.batch_decode(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ )
434
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __A : List[Any] = logging.get_logger(__name__) __A : Any = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] __A : Optional[int] = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" A__ : Union[str, Any] =torch.load(UpperCamelCase , map_location="cpu" ) return sd def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : int=rename_keys_prefix ): """simple docstring""" A__ : List[str] =OrderedDict() A__ : str =torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue A__ : Optional[Any] =key for name_pair in rename_keys_prefix: A__ : int =new_key.replace(name_pair[0] , name_pair[1] ) A__ : Dict =d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately A__ : Optional[int] =new_d["cls.predictions.bias"] return new_d @torch.no_grad() def lowercase ( UpperCamelCase : Dict , UpperCamelCase : List[str] ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: A__ : Any ="pretraining" if "vcr" in checkpoint_path: A__ : Union[str, Any] ={"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: A__ : Optional[Any] ={"visual_embedding_dim": 2048} elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048} elif "nlvr" in checkpoint_path: A__ : List[str] ={"visual_embedding_dim": 1024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 512} A__ : List[str] ="multichoice" elif "vqa_advanced" in checkpoint_path: A__ : Any ={"visual_embedding_dim": 2048} A__ : str ="vqa_advanced" elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048, "num_labels": 3129} A__ : str ="vqa" elif "nlvr" in checkpoint_path: A__ : str ={ "visual_embedding_dim": 1024, "num_labels": 2, } A__ : Dict ="nlvr" A__ : Union[str, Any] =VisualBertConfig(**UpperCamelCase ) # Load State Dict A__ : int =load_state_dict(UpperCamelCase ) A__ : Tuple =get_new_dict(UpperCamelCase , UpperCamelCase ) if model_type == "pretraining": A__ : str =VisualBertForPreTraining(UpperCamelCase ) elif model_type == "vqa": A__ : Optional[int] =VisualBertForQuestionAnswering(UpperCamelCase ) elif model_type == "nlvr": A__ : Union[str, Any] =VisualBertForVisualReasoning(UpperCamelCase ) elif model_type == "multichoice": A__ : Union[str, Any] =VisualBertForMultipleChoice(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) # Save Checkpoints Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": __A : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("orig_checkpoint_path", type=str, help="A path to .th on local filesystem.") parser.add_argument("pytorch_dump_folder_path", type=str, help="Path to the output PyTorch model.") __A : str = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
656
0
import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def __a ( A__ : List[Any] , A__ : Optional[int] ): SCREAMING_SNAKE_CASE = XCLIPTextConfig() # derive patch size from model name SCREAMING_SNAKE_CASE = model_name.find("patch" ) SCREAMING_SNAKE_CASE = int(model_name[start_idx + len("patch" ) : start_idx + len("patch" ) + 2] ) SCREAMING_SNAKE_CASE = XCLIPVisionConfig(patch_size=A__ , num_frames=A__ ) if "large" in model_name: SCREAMING_SNAKE_CASE = 768 SCREAMING_SNAKE_CASE = 3072 SCREAMING_SNAKE_CASE = 12 SCREAMING_SNAKE_CASE = 1024 SCREAMING_SNAKE_CASE = 4096 SCREAMING_SNAKE_CASE = 16 SCREAMING_SNAKE_CASE = 24 SCREAMING_SNAKE_CASE = 768 SCREAMING_SNAKE_CASE = 3072 if model_name == "xclip-large-patch14-16-frames": SCREAMING_SNAKE_CASE = 336 SCREAMING_SNAKE_CASE = XCLIPConfig.from_text_vision_configs(A__ , A__ ) if "large" in model_name: SCREAMING_SNAKE_CASE = 768 return config def __a ( A__ : List[str] ): # text encoder if name == "token_embedding.weight": SCREAMING_SNAKE_CASE = name.replace("token_embedding.weight" , "text_model.embeddings.token_embedding.weight" ) if name == "positional_embedding": SCREAMING_SNAKE_CASE = name.replace("positional_embedding" , "text_model.embeddings.position_embedding.weight" ) if "ln_1" in name: SCREAMING_SNAKE_CASE = name.replace("ln_1" , "layer_norm1" ) if "ln_2" in name: SCREAMING_SNAKE_CASE = name.replace("ln_2" , "layer_norm2" ) if "c_fc" in name: SCREAMING_SNAKE_CASE = name.replace("c_fc" , "fc1" ) if "c_proj" in name: SCREAMING_SNAKE_CASE = name.replace("c_proj" , "fc2" ) if name.startswith("transformer.resblocks" ): SCREAMING_SNAKE_CASE = name.replace("transformer.resblocks" , "text_model.encoder.layers" ) if "attn.out_proj" in name and "message" not in name: SCREAMING_SNAKE_CASE = name.replace("attn.out_proj" , "self_attn.out_proj" ) if "ln_final" in name: SCREAMING_SNAKE_CASE = name.replace("ln_final" , "text_model.final_layer_norm" ) # visual encoder if name == "visual.class_embedding": SCREAMING_SNAKE_CASE = name.replace("visual.class_embedding" , "vision_model.embeddings.class_embedding" ) if name == "visual.positional_embedding": SCREAMING_SNAKE_CASE = name.replace("visual.positional_embedding" , "vision_model.embeddings.position_embedding.weight" ) if name.startswith("visual.transformer.resblocks" ): SCREAMING_SNAKE_CASE = name.replace("visual.transformer.resblocks" , "vision_model.encoder.layers" ) if "visual.conv1" in name: SCREAMING_SNAKE_CASE = name.replace("visual.conv1" , "vision_model.embeddings.patch_embedding" ) if "visual.ln_pre" in name: SCREAMING_SNAKE_CASE = name.replace("visual.ln_pre" , "vision_model.pre_layernorm" ) if "visual.ln_post" in name: SCREAMING_SNAKE_CASE = name.replace("visual.ln_post" , "vision_model.post_layernorm" ) if "visual.proj" in name: SCREAMING_SNAKE_CASE = name.replace("visual.proj" , "visual_projection.weight" ) if "text_projection" in name: SCREAMING_SNAKE_CASE = name.replace("text_projection" , "text_projection.weight" ) # things on top if "prompts_visual_proj" in name: SCREAMING_SNAKE_CASE = name.replace("prompts_visual_proj" , "prompts_visual_projection" ) if "prompts_visual_ln" in name: SCREAMING_SNAKE_CASE = name.replace("prompts_visual_ln" , "prompts_visual_layernorm" ) # mit if name == "mit.positional_embedding": SCREAMING_SNAKE_CASE = name.replace("positional" , "position" ) if name.startswith("mit.resblocks" ): SCREAMING_SNAKE_CASE = name.replace("mit.resblocks" , "mit.encoder.layers" ) # prompts generator if name.startswith("prompts_generator.norm" ): SCREAMING_SNAKE_CASE = name.replace("prompts_generator.norm" , "prompts_generator.layernorm" ) return name def __a ( A__ : int , A__ : Union[str, Any] ): for key in orig_state_dict.copy().keys(): SCREAMING_SNAKE_CASE = orig_state_dict.pop(A__ ) if "attn.in_proj" in key: SCREAMING_SNAKE_CASE = key.split("." ) if key.startswith("visual" ): SCREAMING_SNAKE_CASE = key_split[3] SCREAMING_SNAKE_CASE = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: SCREAMING_SNAKE_CASE = val[ :dim, : ] SCREAMING_SNAKE_CASE = val[ dim : dim * 2, : ] SCREAMING_SNAKE_CASE = val[ -dim:, : ] else: SCREAMING_SNAKE_CASE = val[ :dim ] SCREAMING_SNAKE_CASE = val[ dim : dim * 2 ] SCREAMING_SNAKE_CASE = val[ -dim: ] else: if "weight" in key: SCREAMING_SNAKE_CASE = val[ :dim, : ] SCREAMING_SNAKE_CASE = val[ dim : dim * 2, : ] SCREAMING_SNAKE_CASE = val[ -dim:, : ] else: SCREAMING_SNAKE_CASE = val[:dim] SCREAMING_SNAKE_CASE = val[ dim : dim * 2 ] SCREAMING_SNAKE_CASE = val[-dim:] elif key.startswith("mit" ): SCREAMING_SNAKE_CASE = key_split[2] SCREAMING_SNAKE_CASE = config.vision_config.mit_hidden_size if "weight" in key: SCREAMING_SNAKE_CASE = val[:dim, :] SCREAMING_SNAKE_CASE = val[dim : dim * 2, :] SCREAMING_SNAKE_CASE = val[-dim:, :] else: SCREAMING_SNAKE_CASE = val[:dim] SCREAMING_SNAKE_CASE = val[dim : dim * 2] SCREAMING_SNAKE_CASE = val[-dim:] else: SCREAMING_SNAKE_CASE = key_split[2] SCREAMING_SNAKE_CASE = config.text_config.hidden_size if "weight" in key: SCREAMING_SNAKE_CASE = val[:dim, :] SCREAMING_SNAKE_CASE = val[ dim : dim * 2, : ] SCREAMING_SNAKE_CASE = val[-dim:, :] else: SCREAMING_SNAKE_CASE = val[:dim] SCREAMING_SNAKE_CASE = val[ dim : dim * 2 ] SCREAMING_SNAKE_CASE = val[-dim:] else: SCREAMING_SNAKE_CASE = rename_key(A__ ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: SCREAMING_SNAKE_CASE = val.T SCREAMING_SNAKE_CASE = val return orig_state_dict def __a ( A__ : Optional[int] ): if num_frames == 8: SCREAMING_SNAKE_CASE = "eating_spaghetti_8_frames.npy" elif num_frames == 16: SCREAMING_SNAKE_CASE = "eating_spaghetti.npy" elif num_frames == 32: SCREAMING_SNAKE_CASE = "eating_spaghetti_32_frames.npy" SCREAMING_SNAKE_CASE = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename=A__ , repo_type="dataset" , ) SCREAMING_SNAKE_CASE = np.load(A__ ) return list(A__ ) def __a ( A__ : List[str] , A__ : List[Any]=None , A__ : int=False ): SCREAMING_SNAKE_CASE = { # fully supervised kinetics-400 checkpoints "xclip-base-patch32": "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth", "xclip-base-patch32-16-frames": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth" ), "xclip-base-patch16": "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth", "xclip-base-patch16-16-frames": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth" ), "xclip-large-patch14": "https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb", "xclip-large-patch14-16-frames": "https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f", # fully supervised kinetics-600 checkpoints "xclip-base-patch16-kinetics-600": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth" ), "xclip-base-patch16-kinetics-600-16-frames": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth" ), "xclip-large-patch14-kinetics-600": "https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be", # few shot "xclip-base-patch16-hmdb-2-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth" ), "xclip-base-patch16-hmdb-4-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth" ), "xclip-base-patch16-hmdb-8-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth" ), "xclip-base-patch16-hmdb-16-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth" ), "xclip-base-patch16-ucf-2-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth" ), "xclip-base-patch16-ucf-4-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth" ), "xclip-base-patch16-ucf-8-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth" ), "xclip-base-patch16-ucf-16-shot": ( "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth" ), # zero shot "xclip-base-patch16-zero-shot": "https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth", } SCREAMING_SNAKE_CASE = model_to_url[model_name] SCREAMING_SNAKE_CASE = 8 if "16-frames" in model_name: SCREAMING_SNAKE_CASE = 16 elif "shot" in model_name: SCREAMING_SNAKE_CASE = 32 SCREAMING_SNAKE_CASE = get_xclip_config(A__ , A__ ) SCREAMING_SNAKE_CASE = XCLIPModel(A__ ) model.eval() if "drive" in checkpoint_url: SCREAMING_SNAKE_CASE = "pytorch_model.bin" gdown.cached_download(A__ , A__ , quiet=A__ ) SCREAMING_SNAKE_CASE = torch.load(A__ , map_location="cpu" )["model"] else: SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(A__ )["model"] SCREAMING_SNAKE_CASE = convert_state_dict(A__ , A__ ) SCREAMING_SNAKE_CASE = XCLIPModel(A__ ) SCREAMING_SNAKE_CASE = model.load_state_dict(A__ , strict=A__ ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() SCREAMING_SNAKE_CASE = 336 if model_name == "xclip-large-patch14-16-frames" else 224 SCREAMING_SNAKE_CASE = VideoMAEImageProcessor(size=A__ ) SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("openai/clip-vit-base-patch32" ) SCREAMING_SNAKE_CASE = CLIPTokenizerFast.from_pretrained("openai/clip-vit-base-patch32" ) SCREAMING_SNAKE_CASE = XCLIPProcessor(image_processor=A__ , tokenizer=A__ ) SCREAMING_SNAKE_CASE = prepare_video(A__ ) SCREAMING_SNAKE_CASE = processor( text=["playing sports", "eating spaghetti", "go shopping"] , videos=A__ , return_tensors="pt" , padding=A__ ) print("Shape of pixel values:" , inputs.pixel_values.shape ) with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**A__ ) # Verify outputs SCREAMING_SNAKE_CASE = outputs.logits_per_video SCREAMING_SNAKE_CASE = logits_per_video.softmax(dim=1 ) print("Probs:" , A__ ) # kinetics-400 if model_name == "xclip-base-patch32": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_0_1_9, 0.9_9_5_1, 0.0_0_3_0]] ) elif model_name == "xclip-base-patch32-16-frames": SCREAMING_SNAKE_CASE = torch.tensor([[7.0_9_9_9E-0_4, 9.9_8_8_3E-0_1, 4.5_5_8_0E-0_4]] ) elif model_name == "xclip-base-patch16": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_0_8_3, 0.9_6_8_1, 0.0_2_3_6]] ) elif model_name == "xclip-base-patch16-16-frames": SCREAMING_SNAKE_CASE = torch.tensor([[7.6_9_3_7E-0_4, 9.9_7_2_8E-0_1, 1.9_4_7_3E-0_3]] ) elif model_name == "xclip-large-patch14": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_0_6_2, 0.9_8_6_4, 0.0_0_7_5]] ) elif model_name == "xclip-large-patch14-16-frames": SCREAMING_SNAKE_CASE = torch.tensor([[3.3_8_7_7E-0_4, 9.9_9_3_7E-0_1, 2.8_8_8_8E-0_4]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_5_5_5, 0.8_9_1_4, 0.0_5_3_1]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": SCREAMING_SNAKE_CASE = torch.tensor([[3.8_5_5_4E-0_4, 9.9_9_2_9E-0_1, 3.2_7_5_4E-0_4]] ) elif model_name == "xclip-large-patch14-kinetics-600": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_0_3_6, 0.9_9_2_0, 0.0_0_4_5]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": SCREAMING_SNAKE_CASE = torch.tensor([[7.1_8_9_0E-0_6, 9.9_9_9_4E-0_1, 5.6_5_5_9E-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": SCREAMING_SNAKE_CASE = torch.tensor([[1.0_3_2_0E-0_5, 9.9_9_9_3E-0_1, 6.2_4_3_5E-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": SCREAMING_SNAKE_CASE = torch.tensor([[4.1_3_7_7E-0_6, 9.9_9_9_0E-0_1, 9.8_3_8_6E-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": SCREAMING_SNAKE_CASE = torch.tensor([[4.1_3_4_7E-0_5, 9.9_9_6_2E-0_1, 3.3_4_1_1E-0_4]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": SCREAMING_SNAKE_CASE = torch.tensor([[8.5_8_5_7E-0_5, 9.9_9_2_8E-0_1, 6.3_2_9_1E-0_4]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": SCREAMING_SNAKE_CASE = torch.tensor([[8.5_8_5_7E-0_5, 9.9_9_2_8E-0_1, 6.3_2_9_1E-0_4]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": SCREAMING_SNAKE_CASE = torch.tensor([[0.0_0_2_7, 0.9_9_0_4, 0.0_0_7_0]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": SCREAMING_SNAKE_CASE = torch.tensor([[9.8_2_1_9E-0_4, 9.9_5_9_3E-0_1, 3.0_8_6_3E-0_3]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": SCREAMING_SNAKE_CASE = torch.tensor([[3.5_0_8_2E-0_4, 9.9_7_8_5E-0_1, 1.7_9_6_6E-0_3]] ) else: raise ValueError(F"Model name {model_name} not supported" ) assert torch.allclose(A__ , A__ , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) if push_to_hub: print("Pushing model, processor and slow tokenizer files to the hub..." ) model.push_to_hub(A__ , organization="nielsr" ) processor.push_to_hub(A__ , organization="nielsr" ) slow_tokenizer.push_to_hub(A__ , organization="nielsr" ) if __name__ == "__main__": __A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='xclip-base-patch32', type=str, help='Name of the model.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __A : int = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
16
"""simple docstring""" __A : Union[str, Any] = {str(digit): digit**5 for digit in range(10)} def lowercase ( UpperCamelCase : int ): """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase ) ) def lowercase ( ): """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(UpperCamelCase ) ) if __name__ == "__main__": print(solution())
656
0
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class A__ ( _UpperCamelCase , unittest.TestCase ): pass @nightly @require_onnxruntime @require_torch_gpu class A__ ( unittest.TestCase ): @property def __UpperCamelCase ( self : Optional[int] ) -> Any: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __UpperCamelCase ( self : int ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =ort.SessionOptions() _SCREAMING_SNAKE_CASE =False return options def __UpperCamelCase ( self : Dict ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo.png''' ) _SCREAMING_SNAKE_CASE =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo_mask.png''' ) _SCREAMING_SNAKE_CASE =OnnxStableDiffusionInpaintPipeline.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , revision='''onnx''' , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE ="A red cat sitting on a park bench" _SCREAMING_SNAKE_CASE =np.random.RandomState(0 ) _SCREAMING_SNAKE_CASE =pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , mask_image=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type='''np''' , ) _SCREAMING_SNAKE_CASE =output.images _SCREAMING_SNAKE_CASE =images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.25_14, 0.30_07, 0.35_17, 0.17_90, 0.23_82, 0.31_67, 0.19_44, 0.22_73, 0.24_64] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __UpperCamelCase ( self : Optional[Any] ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo.png''' ) _SCREAMING_SNAKE_CASE =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo_mask.png''' ) _SCREAMING_SNAKE_CASE =LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , subfolder='''scheduler''' , revision='''onnx''' ) _SCREAMING_SNAKE_CASE =OnnxStableDiffusionInpaintPipeline.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , revision='''onnx''' , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE ="A red cat sitting on a park bench" _SCREAMING_SNAKE_CASE =np.random.RandomState(0 ) _SCREAMING_SNAKE_CASE =pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , mask_image=UpperCamelCase__ , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCamelCase__ , output_type='''np''' , ) _SCREAMING_SNAKE_CASE =output.images _SCREAMING_SNAKE_CASE =images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _SCREAMING_SNAKE_CASE =np.array([0.00_86, 0.00_77, 0.00_83, 0.00_93, 0.01_07, 0.01_39, 0.00_94, 0.00_97, 0.01_25] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
691
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig __A : Optional[Any] = logging.get_logger(__name__) # General docstring __A : str = "PoolFormerConfig" # Base docstring __A : Optional[Any] = "sail/poolformer_s12" __A : List[Any] = [1, 512, 7, 7] # Image classification docstring __A : List[str] = "sail/poolformer_s12" __A : Tuple = "tabby, tabby cat" __A : Tuple = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def lowercase ( UpperCamelCase : Any , UpperCamelCase : float = 0.0 , UpperCamelCase : bool = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input A__ : Tuple =1 - drop_prob A__ : List[str] =(input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets A__ : Any =keep_prob + torch.rand(UpperCamelCase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize A__ : Optional[int] =input.div(UpperCamelCase ) * random_tensor return output class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : Optional[float] = None ): super().__init__() A__ : Optional[int] =drop_prob def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : torch.Tensor ): return drop_path(UpperCamelCase__ , self.drop_prob , self.training ) def _UpperCAmelCase ( self : List[str] ): return "p={}".format(self.drop_prob ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): super().__init__() A__ : Optional[int] =patch_size if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (patch_size, patch_size) A__ : Optional[int] =stride if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (stride, stride) A__ : int =padding if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (padding, padding) A__ : Any =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=UpperCamelCase__ ) A__ : Any =norm_layer(UpperCamelCase__ ) if norm_layer else nn.Identity() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : str ): A__ : List[str] =self.projection(UpperCamelCase__ ) A__ : Any =self.norm(UpperCamelCase__ ) return embeddings class __lowerCAmelCase ( nn.GroupNorm): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Dict , **UpperCamelCase__ : Union[str, Any] ): super().__init__(1 , UpperCamelCase__ , **UpperCamelCase__ ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Optional[int] ): super().__init__() A__ : Any =nn.AvgPoolad(UpperCamelCase__ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[str] ): return self.pool(UpperCamelCase__ ) - hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] ): super().__init__() A__ : List[Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Union[str, Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Dict =PoolFormerDropPath(UpperCamelCase__ ) if isinstance(config.hidden_act , UpperCamelCase__ ): A__ : Tuple =ACTaFN[config.hidden_act] else: A__ : Optional[Any] =config.hidden_act def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict ): A__ : Optional[Any] =self.conva(UpperCamelCase__ ) A__ : List[str] =self.act_fn(UpperCamelCase__ ) A__ : List[str] =self.drop(UpperCamelCase__ ) A__ : Optional[int] =self.conva(UpperCamelCase__ ) A__ : Optional[Any] =self.drop(UpperCamelCase__ ) return hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Any ): super().__init__() A__ : Optional[int] =PoolFormerPooling(UpperCamelCase__ ) A__ : List[str] =PoolFormerOutput(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) # Useful for training neural nets A__ : Tuple =PoolFormerDropPath(UpperCamelCase__ ) if drop_path > 0.0 else nn.Identity() A__ : Optional[Any] =config.use_layer_scale if config.use_layer_scale: A__ : List[str] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) A__ : List[Any] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : Optional[int] ): if self.use_layer_scale: A__ : Optional[int] =self.pooling(self.before_norm(UpperCamelCase__ ) ) A__ : Union[str, Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection A__ : Union[str, Any] =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : Tuple =() A__ : List[str] =self.output(self.after_norm(UpperCamelCase__ ) ) A__ : Optional[Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection A__ : str =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : List[Any] =(output,) + outputs return outputs else: A__ : Tuple =self.drop_path(self.pooling(self.before_norm(UpperCamelCase__ ) ) ) # First residual connection A__ : Optional[Any] =pooling_output + hidden_states A__ : Tuple =() # Second residual connection inside the PoolFormerOutput block A__ : List[str] =self.drop_path(self.output(self.after_norm(UpperCamelCase__ ) ) ) A__ : Any =hidden_states + layer_output A__ : Tuple =(output,) + outputs return outputs class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : List[str] ): super().__init__() A__ : Tuple =config # stochastic depth decay rule A__ : Dict =[x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings A__ : Tuple =[] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) A__ : List[str] =nn.ModuleList(UpperCamelCase__ ) # Transformer blocks A__ : Union[str, Any] =[] A__ : Any =0 for i in range(config.num_encoder_blocks ): # each block consists of layers A__ : Union[str, Any] =[] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCamelCase__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCamelCase__ ) ) A__ : str =nn.ModuleList(UpperCamelCase__ ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : Optional[int]=True ): A__ : Union[str, Any] =() if output_hidden_states else None A__ : Dict =pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): A__ , A__ : List[Any] =layers # Get patch embeddings from hidden_states A__ : Any =embedding_layer(UpperCamelCase__ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCamelCase__ ): A__ : List[str] =blk(UpperCamelCase__ ) A__ : Tuple =layer_outputs[0] if output_hidden_states: A__ : List[Any] =all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : List[str] = PoolFormerConfig __magic_name__ : int = """poolformer""" __magic_name__ : Any = """pixel_values""" __magic_name__ : Any = True def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : str ): if isinstance(UpperCamelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any]=False ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A__ : Optional[Any] =value __A : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : Dict ): super().__init__(UpperCamelCase__ ) A__ : List[Any] =config A__ : Optional[Any] =PoolFormerEncoder(UpperCamelCase__ ) # Initialize weights and apply final processing self.post_init() def _UpperCAmelCase ( self : Tuple ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : int =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A__ : Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) A__ : List[Any] =self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : int =encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[Any] ): super().__init__() A__ : List[str] =nn.Linear(config.hidden_size , config.hidden_size ) def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] ): A__ : int =self.dense(UpperCamelCase__ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : str ): super().__init__(UpperCamelCase__ ) A__ : List[str] =config.num_labels A__ : Optional[int] =PoolFormerModel(UpperCamelCase__ ) # Final norm A__ : Dict =PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head A__ : Dict =( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.LongTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : Tuple =return_dict if return_dict is not None else self.config.use_return_dict A__ : List[str] =self.poolformer( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : str =outputs[0] A__ : List[Any] =self.classifier(self.norm(UpperCamelCase__ ).mean([-2, -1] ) ) A__ : Optional[Any] =None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: A__ : int ="regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): A__ : Tuple ="single_label_classification" else: A__ : Optional[int] ="multi_label_classification" if self.config.problem_type == "regression": A__ : Dict =MSELoss() if self.num_labels == 1: A__ : Optional[Any] =loss_fct(logits.squeeze() , labels.squeeze() ) else: A__ : List[str] =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": A__ : Tuple =CrossEntropyLoss() A__ : int =loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": A__ : List[Any] =BCEWithLogitsLoss() A__ : str =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: A__ : Optional[int] =(logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states )
656
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): _lowerCamelCase : str = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right _lowerCamelCase : Dict = 128_022 _lowerCamelCase : Optional[Any] = 128_028 @require_sentencepiece class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ,unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : str = MaMaaaTokenizer _UpperCAmelCase : str = False _UpperCAmelCase : Union[str, Any] = False _UpperCAmelCase : Any = True def A ( self : List[str] ): '''simple docstring''' super().setUp() _snake_case = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] _snake_case = dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) _snake_case = Path(self.tmpdirname ) save_json(UpperCamelCase__ , save_dir / VOCAB_FILES_NAMES['vocab_file'] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(UpperCamelCase__ , save_dir / VOCAB_FILES_NAMES['spm_file'] ) _snake_case = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : List[str] , **lowercase : Optional[Any] ): '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def A ( self : Optional[int] , lowercase : Optional[int] ): '''simple docstring''' return ( "This is a test", "This is a test", ) def A ( self : Dict ): '''simple docstring''' _snake_case = "</s>" _snake_case = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase__ ) , UpperCamelCase__ ) def A ( self : str ): '''simple docstring''' _snake_case = self.get_tokenizer() _snake_case = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '</s>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '<s>' ) self.assertEqual(len(UpperCamelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip('Skip this test while all models are still to be uploaded.' ) def A ( self : List[Any] ): '''simple docstring''' pass def A ( self : Tuple ): '''simple docstring''' _snake_case = self.get_tokenizer() _snake_case = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCamelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [2, 3, 4, 5, 6] , ) _snake_case = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(UpperCamelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) _snake_case = tokenizer.convert_tokens_to_string(UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , 'This is a test' ) @slow def A ( self : int ): '''simple docstring''' _snake_case = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase__ , model_name='facebook/m2m100_418M' , revision='c168bae485c864188cf9aa0e4108b0b6934dc91e' , ) @require_torch @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = """facebook/m2m100_418M""" _UpperCAmelCase : List[Any] = [ """In my opinion, there are two levels of response from the French government.""", """NSA Affair Emphasizes Complete Lack of Debate on Intelligence""", ] _UpperCAmelCase : Optional[int] = [ """Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.""", """L'affaire NSA souligne l'absence totale de débat sur le renseignement""", ] # fmt: off _UpperCAmelCase : List[str] = [EN_CODE, 5_9_3, 1_9_4_9, 1_1_5_7_8_1, 4, 7_1_5_8_6, 4_2_3_4, 6_0_6_3_3, 1_2_6_2_3_3, 4_3_2, 1_2_3_8_0_8, 1_5_5_9_2, 1_1_9_7, 1_1_7_1_3_2, 1_2_0_6_1_8, 5, 2] @classmethod def A ( cls : List[Any] ): '''simple docstring''' _snake_case = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en' , tgt_lang='fr' ) _snake_case = 1 return cls def A ( self : Optional[Any] ): '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id('ar' ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id('en' ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id('ro' ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id('mr' ) , 128_063 ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = self.tokenizer.get_vocab() self.assertEqual(len(UpperCamelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab['<unk>'] , 3 ) self.assertIn(self.tokenizer.get_lang_token('en' ) , UpperCamelCase__ ) def A ( self : Union[str, Any] ): '''simple docstring''' _snake_case = "en" _snake_case = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase__ ) def A ( self : List[Any] ): '''simple docstring''' self.assertIn(UpperCamelCase__ , self.tokenizer.all_special_ids ) # fmt: off _snake_case = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on _snake_case = self.tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) _snake_case = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase__ ) def A ( self : Optional[int] ): '''simple docstring''' _snake_case = tempfile.mkdtemp() _snake_case = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(UpperCamelCase__ ) _snake_case = MaMaaaTokenizer.from_pretrained(UpperCamelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , UpperCamelCase__ ) @require_torch def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = "en" _snake_case = "fr" _snake_case = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase__ , return_tensors='pt' ) _snake_case = shift_tokens_right( batch['labels'] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _snake_case = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('mr' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _snake_case = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('zh' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def A ( self : int ): '''simple docstring''' _snake_case = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('mr' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _snake_case = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('zh' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def A ( self : Dict ): '''simple docstring''' _snake_case = self.tokenizer._build_translation_inputs('A test' , return_tensors='pt' , src_lang='en' , tgt_lang='ar' ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , { # en_XX, A, test, EOS 'input_ids': [[128_022, 58, 4_183, 2]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 128_006, } , )
686
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : int = IFInpaintingSuperResolutionPipeline __magic_name__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __magic_name__ : str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""}) __magic_name__ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def _UpperCAmelCase ( self : Union[str, Any] ): return self._get_superresolution_dummy_components() def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[int]=0 ): if str(UpperCamelCase__ ).startswith("mps" ): A__ : Any =torch.manual_seed(UpperCamelCase__ ) else: A__ : Dict =torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A__ : Tuple =floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Optional[int] =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Any =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : List[str] ={ "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _UpperCAmelCase ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _UpperCAmelCase ( self : int ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _UpperCAmelCase ( self : Tuple ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def _UpperCAmelCase ( self : str ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _UpperCAmelCase ( self : Dict ): self._test_save_load_local() def _UpperCAmelCase ( self : Optional[int] ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
656
0
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): __lowerCAmelCase = torch.nn.Linear(10 , 10 ) __lowerCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) __lowerCAmelCase = Accelerator() __lowerCAmelCase = accelerator.prepare(UpperCamelCase__ ) try: pickle.loads(pickle.dumps(UpperCamelCase__ ) ) except Exception as e: self.fail(f"Accelerated optimizer pickling failed with {e}" ) AcceleratorState._reset_state()
636
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __A : Any = { "configuration_efficientformer": [ "EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientFormerConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = ["EfficientFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientFormerForImageClassification", "EfficientFormerForImageClassificationWithTeacher", "EfficientFormerModel", "EfficientFormerPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFEfficientFormerForImageClassification", "TFEfficientFormerForImageClassificationWithTeacher", "TFEfficientFormerModel", "TFEfficientFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys __A : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
"""simple docstring""" import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: float = 0.0 _lowerCamelCase: int = 1 _lowerCamelCase: int = 1 _lowerCamelCase: bool = True _lowerCamelCase: bool = False _lowerCamelCase: bool = False _lowerCamelCase: bool = False _lowerCamelCase: jnp.dtype = jnp.floataa def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> int: A = [] A = [] for i in range(self.num_layers ): A = self.in_channels if i == 0 else self.out_channels A = FlaxResnetBlockaD( in_channels=UpperCamelCase__ ,out_channels=self.out_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) resnets.append(UpperCamelCase__ ) A = FlaxTransformeraDModel( in_channels=self.out_channels ,n_heads=self.num_attention_heads ,d_head=self.out_channels // self.num_attention_heads ,depth=1 ,use_linear_projection=self.use_linear_projection ,only_cross_attention=self.only_cross_attention ,use_memory_efficient_attention=self.use_memory_efficient_attention ,dtype=self.dtype ,) attentions.append(UpperCamelCase__ ) A = resnets A = attentions if self.add_downsample: A = FlaxDownsampleaD(self.out_channels ,dtype=self.dtype ) def __call__( self : List[Any] ,A_ : Any ,A_ : int ,A_ : int ,A_ : List[Any]=True ) -> Dict: A = () for resnet, attn in zip(self.resnets ,self.attentions ): A = resnet(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) A = attn(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) output_states += (hidden_states,) if self.add_downsample: A = self.downsamplers_a(UpperCamelCase__ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: float = 0.0 _lowerCamelCase: int = 1 _lowerCamelCase: bool = True _lowerCamelCase: jnp.dtype = jnp.floataa def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[Any]: A = [] for i in range(self.num_layers ): A = self.in_channels if i == 0 else self.out_channels A = FlaxResnetBlockaD( in_channels=UpperCamelCase__ ,out_channels=self.out_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) resnets.append(UpperCamelCase__ ) A = resnets if self.add_downsample: A = FlaxDownsampleaD(self.out_channels ,dtype=self.dtype ) def __call__( self : Tuple ,A_ : Optional[Any] ,A_ : List[str] ,A_ : List[str]=True ) -> Tuple: A = () for resnet in self.resnets: A = resnet(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) output_states += (hidden_states,) if self.add_downsample: A = self.downsamplers_a(UpperCamelCase__ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: float = 0.0 _lowerCamelCase: int = 1 _lowerCamelCase: int = 1 _lowerCamelCase: bool = True _lowerCamelCase: bool = False _lowerCamelCase: bool = False _lowerCamelCase: bool = False _lowerCamelCase: jnp.dtype = jnp.floataa def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Optional[Any]: A = [] A = [] for i in range(self.num_layers ): A = self.in_channels if (i == self.num_layers - 1) else self.out_channels A = self.prev_output_channel if i == 0 else self.out_channels A = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels ,out_channels=self.out_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) resnets.append(UpperCamelCase__ ) A = FlaxTransformeraDModel( in_channels=self.out_channels ,n_heads=self.num_attention_heads ,d_head=self.out_channels // self.num_attention_heads ,depth=1 ,use_linear_projection=self.use_linear_projection ,only_cross_attention=self.only_cross_attention ,use_memory_efficient_attention=self.use_memory_efficient_attention ,dtype=self.dtype ,) attentions.append(UpperCamelCase__ ) A = resnets A = attentions if self.add_upsample: A = FlaxUpsampleaD(self.out_channels ,dtype=self.dtype ) def __call__( self : Tuple ,A_ : Any ,A_ : Optional[int] ,A_ : Union[str, Any] ,A_ : Any ,A_ : Optional[Any]=True ) -> Dict: for resnet, attn in zip(self.resnets ,self.attentions ): # pop res hidden states A = res_hidden_states_tuple[-1] A = res_hidden_states_tuple[:-1] A = jnp.concatenate((hidden_states, res_hidden_states) ,axis=-1 ) A = resnet(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) A = attn(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) if self.add_upsample: A = self.upsamplers_a(UpperCamelCase__ ) return hidden_states class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: int _lowerCamelCase: float = 0.0 _lowerCamelCase: int = 1 _lowerCamelCase: bool = True _lowerCamelCase: jnp.dtype = jnp.floataa def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[str]: A = [] for i in range(self.num_layers ): A = self.in_channels if (i == self.num_layers - 1) else self.out_channels A = self.prev_output_channel if i == 0 else self.out_channels A = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels ,out_channels=self.out_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) resnets.append(UpperCamelCase__ ) A = resnets if self.add_upsample: A = FlaxUpsampleaD(self.out_channels ,dtype=self.dtype ) def __call__( self : str ,A_ : Union[str, Any] ,A_ : Any ,A_ : Tuple ,A_ : Union[str, Any]=True ) -> Union[str, Any]: for resnet in self.resnets: # pop res hidden states A = res_hidden_states_tuple[-1] A = res_hidden_states_tuple[:-1] A = jnp.concatenate((hidden_states, res_hidden_states) ,axis=-1 ) A = resnet(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) if self.add_upsample: A = self.upsamplers_a(UpperCamelCase__ ) return hidden_states class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: float = 0.0 _lowerCamelCase: int = 1 _lowerCamelCase: int = 1 _lowerCamelCase: bool = False _lowerCamelCase: bool = False _lowerCamelCase: jnp.dtype = jnp.floataa def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Union[str, Any]: # there is always at least one resnet A = [ FlaxResnetBlockaD( in_channels=self.in_channels ,out_channels=self.in_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) ] A = [] for _ in range(self.num_layers ): A = FlaxTransformeraDModel( in_channels=self.in_channels ,n_heads=self.num_attention_heads ,d_head=self.in_channels // self.num_attention_heads ,depth=1 ,use_linear_projection=self.use_linear_projection ,use_memory_efficient_attention=self.use_memory_efficient_attention ,dtype=self.dtype ,) attentions.append(UpperCamelCase__ ) A = FlaxResnetBlockaD( in_channels=self.in_channels ,out_channels=self.in_channels ,dropout_prob=self.dropout ,dtype=self.dtype ,) resnets.append(UpperCamelCase__ ) A = resnets A = attentions def __call__( self : str ,A_ : Any ,A_ : str ,A_ : Optional[int] ,A_ : str=True ) -> int: A = self.resnets[0](UpperCamelCase__ ,UpperCamelCase__ ) for attn, resnet in zip(self.attentions ,self.resnets[1:] ): A = attn(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) A = resnet(UpperCamelCase__ ,UpperCamelCase__ ,deterministic=UpperCamelCase__ ) return hidden_states
91
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Tuple =[] for _ in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Dict =[] for step in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] =os.path.join(UpperCamelCase , "schedule.bin" ) torch.save(scheduler.state_dict() , UpperCamelCase ) A__ : Dict =torch.load(UpperCamelCase ) scheduler.load_state_dict(UpperCamelCase ) return lrs @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : Any =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Optional[Any] =torch.tensor([0.4, 0.2, -0.5] ) A__ : Any =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : List[str] =AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(100 ): A__ : Optional[int] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def _UpperCAmelCase ( self : Dict ): A__ : Optional[int] =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Dict =torch.tensor([0.4, 0.2, -0.5] ) A__ : Optional[int] =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : int =Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=UpperCamelCase__ , weight_decay=0.0 , relative_step=UpperCamelCase__ , scale_parameter=UpperCamelCase__ , warmup_init=UpperCamelCase__ , ) for _ in range(1000 ): A__ : List[Any] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' __magic_name__ : Optional[int] = nn.Linear(50 , 50) if is_torch_available() else None __magic_name__ : Any = AdamW(m.parameters() , lr=10.0) if is_torch_available() else None __magic_name__ : Union[str, Any] = 10 def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ , msg=UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[Any] ): A__ : Union[str, Any] ={"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) A__ : Union[str, Any] ={ get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1E-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): A__ , A__ : Any =data A__ : Union[str, Any] =scheduler_func(self.optimizer , **UpperCamelCase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) A__ : int =unwrap_schedule(UpperCamelCase__ , self.num_steps ) self.assertListAlmostEqual( UpperCamelCase__ , UpperCamelCase__ , tol=1E-2 , msg=F'''failed for {scheduler_func} in normal scheduler''' , ) A__ : List[str] =scheduler_func(self.optimizer , **UpperCamelCase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(UpperCamelCase__ ) # wrap to test picklability of the schedule A__ : Tuple =unwrap_and_save_reload_schedule(UpperCamelCase__ , self.num_steps ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ , msg=F'''failed for {scheduler_func} in save and reload''' ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : str ): A__ : int =fn def __call__( self : List[Any] , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : List[Any] ): return self.fn(*UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict ): A__ : str =list(map(self , scheduler.lr_lambdas ) )
656
0
"""simple docstring""" def _UpperCamelCase ( A ): return " ".join( "".join(word[::-1] ) if len(A ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("Hey wollef sroirraw"))
391
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __A : List[Any] = logging.get_logger("transformers.models.speecht5") __A : Optional[Any] = { "speech_encoder_prenet.layer_norm": "speecht5.encoder.prenet.feature_projection.layer_norm", "speech_encoder_prenet.post_extract_proj": "speecht5.encoder.prenet.feature_projection.projection", "speech_encoder_prenet.pos_conv.0": "speecht5.encoder.prenet.pos_conv_embed.conv", "speech_encoder_prenet.mask_emb": "speecht5.encoder.prenet.masked_spec_embed", } __A : Optional[int] = { "text_encoder_prenet.encoder_prenet.0": "speecht5.encoder.prenet.embed_tokens", "text_encoder_prenet.encoder_prenet.1.alpha": "speecht5.encoder.prenet.encode_positions.alpha", } __A : List[str] = { "speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0": "speecht5.decoder.prenet.layers.0", "speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0": "speecht5.decoder.prenet.layers.1", "speech_decoder_prenet.decoder_prenet.0.1": "speecht5.decoder.prenet.final_layer", "speech_decoder_prenet.decoder_prenet.1.alpha": "speecht5.decoder.prenet.encode_positions.alpha", "speech_decoder_prenet.spkembs_layer.0": "speecht5.decoder.prenet.speaker_embeds_layer", } __A : List[Any] = { "speech_decoder_postnet.feat_out": "speech_decoder_postnet.feat_out", "speech_decoder_postnet.prob_out": "speech_decoder_postnet.prob_out", "speech_decoder_postnet.postnet.postnet.0.0": "speech_decoder_postnet.layers.0.conv", "speech_decoder_postnet.postnet.postnet.0.1": "speech_decoder_postnet.layers.0.batch_norm", "speech_decoder_postnet.postnet.postnet.1.0": "speech_decoder_postnet.layers.1.conv", "speech_decoder_postnet.postnet.postnet.1.1": "speech_decoder_postnet.layers.1.batch_norm", "speech_decoder_postnet.postnet.postnet.2.0": "speech_decoder_postnet.layers.2.conv", "speech_decoder_postnet.postnet.postnet.2.1": "speech_decoder_postnet.layers.2.batch_norm", "speech_decoder_postnet.postnet.postnet.3.0": "speech_decoder_postnet.layers.3.conv", "speech_decoder_postnet.postnet.postnet.3.1": "speech_decoder_postnet.layers.3.batch_norm", "speech_decoder_postnet.postnet.postnet.4.0": "speech_decoder_postnet.layers.4.conv", "speech_decoder_postnet.postnet.postnet.4.1": "speech_decoder_postnet.layers.4.batch_norm", } __A : Union[str, Any] = { "text_decoder_prenet.embed_tokens": "speecht5.decoder.prenet.embed_tokens", } __A : Any = { "text_decoder_postnet.output_projection": "text_decoder_postnet.lm_head", } __A : Union[str, Any] = { "encoder.layers.*.self_attn.k_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj", "encoder.layers.*.self_attn.v_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj", "encoder.layers.*.self_attn.q_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj", "encoder.layers.*.self_attn.out_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj", "encoder.layers.*.self_attn_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.layer_norm", "encoder.layers.*.fc1": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense", "encoder.layers.*.fc2": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense", "encoder.layers.*.final_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm", "encoder.layer_norm": "speecht5.encoder.wrapped_encoder.layer_norm", "encoder.pos_emb.pe_k": "speecht5.encoder.wrapped_encoder.embed_positions.pe_k", } __A : Optional[int] = { "decoder.layers.*.self_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj", "decoder.layers.*.self_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj", "decoder.layers.*.self_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj", "decoder.layers.*.self_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj", "decoder.layers.*.self_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm", "decoder.layers.*.encoder_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj", "decoder.layers.*.encoder_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj", "decoder.layers.*.encoder_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj", "decoder.layers.*.encoder_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj", "decoder.layers.*.encoder_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm", "decoder.layers.*.fc1": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense", "decoder.layers.*.fc2": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense", "decoder.layers.*.final_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm", } __A : Union[str, Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __A : Optional[Any] = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : Optional[int] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : int = [] __A : int = [ "encoder.version", "encoder.layers.*.norm_k.weight", "encoder.layers.*.norm_k.bias", "decoder.version", "decoder.layers.*.norm_k.weight", "decoder.layers.*.norm_k.bias", "decoder.pos_emb.pe_k", "speech_encoder_prenet.embed_positions._float_tensor", "text_decoder_prenet.embed_positions._float_tensor", ] __A : Optional[Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "speech_decoder_prenet.*", "speech_decoder_postnet.*", ] __A : Tuple = IGNORE_KEYS + [ "encoder.proj", "speech_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] __A : Union[str, Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] def lowercase ( UpperCamelCase : str , UpperCamelCase : Union[str, Any] , UpperCamelCase : int , UpperCamelCase : List[Any] , UpperCamelCase : int ): """simple docstring""" for attribute in key.split("." ): A__ : Dict =getattr(UpperCamelCase , UpperCamelCase ) if weight_type is not None: A__ : Union[str, Any] =getattr(UpperCamelCase , UpperCamelCase ).shape else: A__ : Tuple =hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": A__ : Any =value elif weight_type == "weight_g": A__ : Any =value elif weight_type == "weight_v": A__ : Any =value elif weight_type == "bias": A__ : Tuple =value elif weight_type == "running_mean": A__ : Dict =value elif weight_type == "running_var": A__ : List[str] =value elif weight_type == "num_batches_tracked": A__ : Dict =value else: A__ : Optional[int] =value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''' ) def lowercase ( UpperCamelCase : Tuple , UpperCamelCase : Tuple ): """simple docstring""" for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: A__ , A__ : List[str] =key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : Dict ): """simple docstring""" A__ : Tuple =[] if task == "s2t": A__ : Dict =hf_model.speechta.encoder.prenet.feature_encoder A__ : int =MAPPING_S2T A__ : List[Any] =IGNORE_KEYS_S2T elif task == "t2s": A__ : Union[str, Any] =None A__ : List[Any] =MAPPING_T2S A__ : Tuple =IGNORE_KEYS_T2S elif task == "s2s": A__ : Optional[Any] =hf_model.speechta.encoder.prenet.feature_encoder A__ : Tuple =MAPPING_S2S A__ : Any =IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''' ) for name, value in fairseq_dict.items(): if should_ignore(UpperCamelCase , UpperCamelCase ): logger.info(F'''{name} was ignored''' ) continue A__ : Optional[Any] =False if "conv_layers" in name: load_conv_layer( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , hf_model.config.feat_extract_norm == "group" , ) A__ : List[Any] =True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: A__ , A__ : Dict =key.split(".*." ) if prefix in name and suffix in name: A__ : int =suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: A__ : List[Any] =True if "*" in mapped_key: A__ : Optional[int] =name.split(UpperCamelCase )[0].split("." )[-2] A__ : int =mapped_key.replace("*" , UpperCamelCase ) if "weight_g" in name: A__ : str ="weight_g" elif "weight_v" in name: A__ : Optional[Any] ="weight_v" elif "bias" in name: A__ : Any ="bias" elif "weight" in name: A__ : Optional[int] ="weight" elif "running_mean" in name: A__ : Tuple ="running_mean" elif "running_var" in name: A__ : Optional[int] ="running_var" elif "num_batches_tracked" in name: A__ : str ="num_batches_tracked" else: A__ : List[Any] =None set_recursively(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) continue if not is_used: unused_weights.append(UpperCamelCase ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Dict , UpperCamelCase : Any , UpperCamelCase : str , UpperCamelCase : Dict ): """simple docstring""" A__ : Any =full_name.split("conv_layers." )[-1] A__ : Dict =name.split("." ) A__ : int =int(items[0] ) A__ : str =int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) A__ : Optional[Any] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) A__ : Optional[int] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(UpperCamelCase ) @torch.no_grad() def lowercase ( UpperCamelCase : Any , UpperCamelCase : Union[str, Any] , UpperCamelCase : List[str] , UpperCamelCase : str=None , UpperCamelCase : Any=None , UpperCamelCase : Tuple=None , ): """simple docstring""" if config_path is not None: A__ : Any =SpeechTaConfig.from_pretrained(UpperCamelCase ) else: A__ : Any =SpeechTaConfig() if task == "s2t": A__ : Union[str, Any] =config.max_text_positions A__ : Dict =SpeechTaForSpeechToText(UpperCamelCase ) elif task == "t2s": A__ : str =1876 A__ : Optional[int] =600 A__ : Tuple =config.max_speech_positions A__ : Optional[Any] =SpeechTaForTextToSpeech(UpperCamelCase ) elif task == "s2s": A__ : str =1876 A__ : Tuple =config.max_speech_positions A__ : Any =SpeechTaForSpeechToSpeech(UpperCamelCase ) else: raise ValueError(F'''Unknown task name: {task}''' ) if vocab_path: A__ : str =SpeechTaTokenizer(UpperCamelCase , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it A__ : Optional[Any] =AddedToken("<mask>" , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) A__ : int =mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) A__ : Dict =SpeechTaFeatureExtractor() A__ : Tuple =SpeechTaProcessor(tokenizer=UpperCamelCase , feature_extractor=UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) A__ : Union[str, Any] =torch.load(UpperCamelCase ) recursively_load_weights(fairseq_checkpoint["model"] , UpperCamelCase , UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if repo_id: print("Pushing to the hub..." ) processor.push_to_hub(UpperCamelCase ) model.push_to_hub(UpperCamelCase ) if __name__ == "__main__": __A : Dict = argparse.ArgumentParser() parser.add_argument( "--task", default="s2t", type=str, help="Type of the SpeechT5 model you'd like to convert. Should be one of 's2t', 't2s', 's2s'.", ) parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--vocab_path", default=None, type=str, help="Path to SentencePiece model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) __A : str = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
656
0
'''simple docstring''' import argparse import os import re __SCREAMING_SNAKE_CASE : str = "src/transformers" # Pattern that looks at the indentation in a line. __SCREAMING_SNAKE_CASE : Dict = re.compile(r'''^(\s*)\S''') # Pattern that matches `"key":" and puts `key` in group 0. __SCREAMING_SNAKE_CASE : Tuple = re.compile(r'''^\s*\"([^\"]+)\":''') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. __SCREAMING_SNAKE_CASE : List[Any] = re.compile(r'''^\s*_import_structure\[\"([^\"]+)\"\]''') # Pattern that matches `"key",` and puts `key` in group 0. __SCREAMING_SNAKE_CASE : int = re.compile(r'''^\s*\"([^\"]+)\",\s*$''') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. __SCREAMING_SNAKE_CASE : int = re.compile(r'''\[([^\]]+)\]''') def a_ ( UpperCamelCase_ ): A_ = _re_indent.search(UpperCamelCase_ ) return "" if search is None else search.groups()[0] def a_ ( UpperCamelCase_ , UpperCamelCase_="" , UpperCamelCase_=None , UpperCamelCase_=None ): A_ = 0 A_ = code.split("\n" ) if start_prompt is not None: while not lines[index].startswith(UpperCamelCase_ ): index += 1 A_ = ["\n".join(lines[:index] )] else: A_ = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). A_ = [lines[index]] index += 1 while index < len(UpperCamelCase_ ) and (end_prompt is None or not lines[index].startswith(UpperCamelCase_ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(UpperCamelCase_ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + " " ): current_block.append(lines[index] ) blocks.append("\n".join(UpperCamelCase_ ) ) if index < len(UpperCamelCase_ ) - 1: A_ = [lines[index + 1]] index += 1 else: A_ = [] else: blocks.append("\n".join(UpperCamelCase_ ) ) A_ = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(UpperCamelCase_ ) > 0: blocks.append("\n".join(UpperCamelCase_ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(UpperCamelCase_ ): blocks.append("\n".join(lines[index:] ) ) return blocks def a_ ( UpperCamelCase_ ): def _inner(UpperCamelCase_ ): return key(UpperCamelCase_ ).lower().replace("_" , "" ) return _inner def a_ ( UpperCamelCase_ , UpperCamelCase_=None ): # If no key is provided, we use a noop. def noop(UpperCamelCase_ ): return x if key is None: A_ = noop # Constants are all uppercase, they go first. A_ = [obj for obj in objects if key(UpperCamelCase_ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. A_ = [obj for obj in objects if key(UpperCamelCase_ )[0].isupper() and not key(UpperCamelCase_ ).isupper()] # Functions begin with a lowercase, they go last. A_ = [obj for obj in objects if not key(UpperCamelCase_ )[0].isupper()] A_ = ignore_underscore(UpperCamelCase_ ) return sorted(UpperCamelCase_ , key=UpperCamelCase_ ) + sorted(UpperCamelCase_ , key=UpperCamelCase_ ) + sorted(UpperCamelCase_ , key=UpperCamelCase_ ) def a_ ( UpperCamelCase_ ): # This inner function sort imports between [ ]. def _replace(UpperCamelCase_ ): A_ = match.groups()[0] if "," not in imports: return f"[{imports}]" A_ = [part.strip().replace("\"" , "" ) for part in imports.split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: A_ = keys[:-1] return "[" + ", ".join([f"\"{k}\"" for k in sort_objects(UpperCamelCase_ )] ) + "]" A_ = import_statement.split("\n" ) if len(UpperCamelCase_ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. A_ = 2 if lines[1].strip() == "[" else 1 A_ = [(i, _re_strip_line.search(UpperCamelCase_ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] A_ = sort_objects(UpperCamelCase_ , key=lambda UpperCamelCase_ : x[1] ) A_ = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(UpperCamelCase_ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: A_ = _re_bracket_content.sub(_replace , lines[1] ) else: A_ = [part.strip().replace("\"" , "" ) for part in lines[1].split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: A_ = keys[:-1] A_ = get_indent(lines[1] ) + ", ".join([f"\"{k}\"" for k in sort_objects(UpperCamelCase_ )] ) return "\n".join(UpperCamelCase_ ) else: # Finally we have to deal with imports fitting on one line A_ = _re_bracket_content.sub(_replace , UpperCamelCase_ ) return import_statement def a_ ( UpperCamelCase_ , UpperCamelCase_=True ): with open(UpperCamelCase_ , encoding="utf-8" ) as f: A_ = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 A_ = split_code_in_indented_blocks( UpperCamelCase_ , start_prompt="_import_structure = {" , end_prompt="if TYPE_CHECKING:" ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(UpperCamelCase_ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. A_ = main_blocks[block_idx] A_ = block.split("\n" ) # Get to the start of the imports. A_ = 0 while line_idx < len(UpperCamelCase_ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: A_ = len(UpperCamelCase_ ) else: line_idx += 1 if line_idx >= len(UpperCamelCase_ ): continue # Ignore beginning and last line: they don't contain anything. A_ = "\n".join(block_lines[line_idx:-1] ) A_ = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. A_ = split_code_in_indented_blocks(UpperCamelCase_ , indent_level=UpperCamelCase_ ) # We have two categories of import key: list or _import_structure[key].append/extend A_ = _re_direct_key if "_import_structure = {" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. A_ = [(pattern.search(UpperCamelCase_ ).groups()[0] if pattern.search(UpperCamelCase_ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. A_ = [(i, key) for i, key in enumerate(UpperCamelCase_ ) if key is not None] A_ = [x[0] for x in sorted(UpperCamelCase_ , key=lambda UpperCamelCase_ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. A_ = 0 A_ = [] for i in range(len(UpperCamelCase_ ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: A_ = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(UpperCamelCase_ ) count += 1 # And we put our main block back together with its first and last line. A_ = "\n".join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(UpperCamelCase_ ): if check_only: return True else: print(f"Overwriting {file}." ) with open(UpperCamelCase_ , "w" , encoding="utf-8" ) as f: f.write("\n".join(UpperCamelCase_ ) ) def a_ ( UpperCamelCase_=True ): A_ = [] for root, _, files in os.walk(UpperCamelCase_ ): if "__init__.py" in files: A_ = sort_imports(os.path.join(UpperCamelCase_ , "__init__.py" ) , check_only=UpperCamelCase_ ) if result: A_ = [os.path.join(UpperCamelCase_ , "__init__.py" )] if len(UpperCamelCase_ ) > 0: raise ValueError(f"Would overwrite {len(UpperCamelCase_ )} files, run `make style`." ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') __SCREAMING_SNAKE_CASE : List[str] = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
452
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase): '''simple docstring''' __magic_name__ : List[Any] = [R"""h\.\d+\.attn\.bias""", R"""h\.\d+\.attn\.masked_bias"""] @register_to_config def __init__( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : int = 50257 , UpperCamelCase__ : int = 1024 , UpperCamelCase__ : int = 768 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : str = "gelu_new" , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 1E-5 , UpperCamelCase__ : float = 0.02 , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , ): super().__init__() A__ : Dict =prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) A__ : Optional[int] =prefix_inner_dim A__ : Optional[int] =prefix_hidden_dim A__ : Optional[int] =( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : Optional[int] =( nn.Linear(self.prefix_hidden_dim , UpperCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : str =GPTaConfig( vocab_size=UpperCamelCase__ , n_positions=UpperCamelCase__ , n_embd=UpperCamelCase__ , n_layer=UpperCamelCase__ , n_head=UpperCamelCase__ , n_inner=UpperCamelCase__ , activation_function=UpperCamelCase__ , resid_pdrop=UpperCamelCase__ , embd_pdrop=UpperCamelCase__ , attn_pdrop=UpperCamelCase__ , layer_norm_epsilon=UpperCamelCase__ , initializer_range=UpperCamelCase__ , scale_attn_weights=UpperCamelCase__ , use_cache=UpperCamelCase__ , scale_attn_by_inverse_layer_idx=UpperCamelCase__ , reorder_and_upcast_attn=UpperCamelCase__ , ) A__ : Any =GPTaLMHeadModel(UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : Optional[torch.Tensor] = None , UpperCamelCase__ : Optional[torch.Tensor] = None , ): A__ : int =self.transformer.transformer.wte(UpperCamelCase__ ) A__ : Tuple =self.encode_prefix(UpperCamelCase__ ) A__ : Union[str, Any] =self.decode_prefix(UpperCamelCase__ ) A__ : Tuple =torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A__ : Any =self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A__ : List[Any] =torch.cat((dummy_token, input_ids) , dim=1 ) A__ : Any =self.transformer(inputs_embeds=UpperCamelCase__ , labels=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : torch.device ): return torch.zeros(UpperCamelCase__ , self.prefix_length , dtype=torch.intaa , device=UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Tuple ): return self.encode_prefix(UpperCamelCase__ ) @torch.no_grad() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : str ): A__ : Optional[int] =torch.split(UpperCamelCase__ , 1 , dim=0 ) A__ : List[str] =[] A__ : Dict =[] for feature in features: A__ : Any =self.decode_prefix(feature.to(UpperCamelCase__ ) ) # back to the clip feature # Only support beam search for now A__ , A__ : Optional[Any] =self.generate_beam( input_embeds=UpperCamelCase__ , device=UpperCamelCase__ , eos_token_id=UpperCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A__ : Optional[Any] =torch.stack(UpperCamelCase__ ) A__ : Optional[int] =torch.stack(UpperCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : List[str]=None , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : int = 5 , UpperCamelCase__ : int = 67 , UpperCamelCase__ : float = 1.0 , UpperCamelCase__ : Optional[int] = None , ): A__ : str =eos_token_id A__ : Optional[Any] =None A__ : int =None A__ : Union[str, Any] =torch.ones(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.int ) A__ : Any =torch.zeros(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.bool ) if input_embeds is not None: A__ : Union[str, Any] =input_embeds else: A__ : Optional[Any] =self.transformer.transformer.wte(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): A__ : Optional[int] =self.transformer(inputs_embeds=UpperCamelCase__ ) A__ : Tuple =outputs.logits A__ : Union[str, Any] =logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A__ : Optional[Any] =logits.softmax(-1 ).log() if scores is None: A__ , A__ : Union[str, Any] =logits.topk(UpperCamelCase__ , -1 ) A__ : Union[str, Any] =generated.expand(UpperCamelCase__ , *generated.shape[1:] ) A__ , A__ : Optional[int] =next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A__ : str =next_tokens else: A__ : Optional[Any] =tokens.expand(UpperCamelCase__ , *tokens.shape[1:] ) A__ : str =torch.cat((tokens, next_tokens) , dim=1 ) else: A__ : Union[str, Any] =-float(np.inf ) A__ : Dict =0 A__ : Optional[Any] =scores[:, None] + logits seq_lengths[~is_stopped] += 1 A__ : Optional[Any] =scores_sum / seq_lengths[:, None] A__ , A__ : List[Any] =scores_sum_average.view(-1 ).topk(UpperCamelCase__ , -1 ) A__ : Tuple =next_tokens // scores_sum.shape[1] A__ : List[Any] =seq_lengths[next_tokens_source] A__ : int =next_tokens % scores_sum.shape[1] A__ : str =next_tokens.unsqueeze(1 ) A__ : List[Any] =tokens[next_tokens_source] A__ : int =torch.cat((tokens, next_tokens) , dim=1 ) A__ : List[str] =generated[next_tokens_source] A__ : Optional[Any] =scores_sum_average * seq_lengths A__ : Optional[int] =is_stopped[next_tokens_source] A__ : List[str] =self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A__ : str =torch.cat((generated, next_token_embed) , dim=1 ) A__ : str =is_stopped + next_tokens.eq(UpperCamelCase__ ).squeeze() if is_stopped.all(): break A__ : Optional[int] =scores / seq_lengths A__ : List[Any] =scores.argsort(descending=UpperCamelCase__ ) # tokens tensors are already padded to max_seq_length A__ : int =[tokens[i] for i in order] A__ : Any =torch.stack(UpperCamelCase__ , dim=0 ) A__ : int =torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
656
0
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def a_ ( lowercase__ :Union[str, Any], lowercase__ :Union[str, Any]=10 ): __lowerCamelCase = [] for _ in range(lowercase__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def a_ ( lowercase__ :List[str], lowercase__ :Union[str, Any]=10 ): __lowerCamelCase = [] for step in range(lowercase__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: __lowerCamelCase = os.path.join(lowercase__, """schedule.bin""" ) torch.save(scheduler.state_dict(), lowercase__ ) __lowerCamelCase = torch.load(lowercase__ ) scheduler.load_state_dict(lowercase__ ) return lrs @require_torch class __snake_case (unittest.TestCase ): def __a ( self: Dict , A_: List[Any] , A_: Optional[Any] , A_: int ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ ) def __a ( self: Tuple ): __lowerCamelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) __lowerCamelCase = torch.tensor([0.4, 0.2, -0.5] ) __lowerCamelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping __lowerCamelCase = AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(1_00 ): __lowerCamelCase = criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def __a ( self: Dict ): __lowerCamelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) __lowerCamelCase = torch.tensor([0.4, 0.2, -0.5] ) __lowerCamelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping __lowerCamelCase = Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=UpperCamelCase__ , weight_decay=0.0 , relative_step=UpperCamelCase__ , scale_parameter=UpperCamelCase__ , warmup_init=UpperCamelCase__ , ) for _ in range(10_00 ): __lowerCamelCase = criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class __snake_case (unittest.TestCase ): __a = nn.Linear(50 , 50 ) if is_torch_available() else None __a = AdamW(m.parameters() , lr=10.0 ) if is_torch_available() else None __a = 10 def __a ( self: List[Any] , A_: Union[str, Any] , A_: List[str] , A_: Optional[Any] , A_: int=None ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ , msg=UpperCamelCase__ ) def __a ( self: Optional[Any] ): __lowerCamelCase = {"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) __lowerCamelCase = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1E-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): __lowerCamelCase = data __lowerCamelCase = scheduler_func(self.optimizer , **UpperCamelCase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) __lowerCamelCase = unwrap_schedule(UpperCamelCase__ , self.num_steps ) self.assertListAlmostEqual( UpperCamelCase__ , UpperCamelCase__ , tol=1E-2 , msg=f'failed for {scheduler_func} in normal scheduler' , ) __lowerCamelCase = scheduler_func(self.optimizer , **UpperCamelCase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(UpperCamelCase__ ) # wrap to test picklability of the schedule __lowerCamelCase = unwrap_and_save_reload_schedule(UpperCamelCase__ , self.num_steps ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ , msg=f'failed for {scheduler_func} in save and reload' ) class __snake_case : def __init__( self: int , A_: str ): __lowerCamelCase = fn def __call__( self: List[Any] , *A_: Optional[Any] , **A_: List[Any] ): return self.fn(*UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def __a ( self: Dict , A_: Dict ): __lowerCamelCase = list(map(self , scheduler.lr_lambdas ) )
281
"""simple docstring""" import os def lowercase ( ): """simple docstring""" A__ : List[Any] =os.path.dirname(os.path.realpath(UpperCamelCase ) ) A__ : str =os.path.join(UpperCamelCase , "triangle.txt" ) with open(UpperCamelCase ) as f: A__ : Optional[int] =f.readlines() A__ : str =[] for line in triangle: A__ : Union[str, Any] =[] for number in line.strip().split(" " ): numbers_from_line.append(int(UpperCamelCase ) ) a.append(UpperCamelCase ) for i in range(1 , len(UpperCamelCase ) ): for j in range(len(a[i] ) ): A__ : Union[str, Any] =a[i - 1][j] if j != len(a[i - 1] ) else 0 A__ : Union[str, Any] =a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(UpperCamelCase , UpperCamelCase ) return max(a[-1] ) if __name__ == "__main__": print(solution())
656
0
_UpperCamelCase : List[Any] =[ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] _UpperCamelCase : Optional[Any] =[ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] _UpperCamelCase : Any =[ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] _UpperCamelCase : Tuple =[ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] _UpperCamelCase : Optional[int] =[ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] _UpperCamelCase : Optional[int] =[ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] _UpperCamelCase : List[str] =[ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] _UpperCamelCase : List[Any] =[ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
206
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A : int = logging.get_logger(__name__) def lowercase ( UpperCamelCase : Any ): """simple docstring""" A__ : str =OrderedDict() for key, value in state_dict.items(): if key.startswith("module.encoder" ): A__ : Dict =key.replace("module.encoder" , "glpn.encoder" ) if key.startswith("module.decoder" ): A__ : Optional[int] =key.replace("module.decoder" , "decoder.stages" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 A__ : Tuple =key[key.find("patch_embed" ) + len("patch_embed" )] A__ : Optional[Any] =key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(UpperCamelCase )-1}''' ) if "norm" in key: A__ : Dict =key.replace("norm" , "layer_norm" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 A__ : Any =key[key.find("glpn.encoder.layer_norm" ) + len("glpn.encoder.layer_norm" )] A__ : Tuple =key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(UpperCamelCase )-1}''' ) if "layer_norm1" in key: A__ : List[Any] =key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: A__ : Optional[int] =key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 A__ : int =key[key.find("block" ) + len("block" )] A__ : Optional[Any] =key.replace(F'''block{idx}''' , F'''block.{int(UpperCamelCase )-1}''' ) if "attn.q" in key: A__ : Optional[Any] =key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: A__ : Union[str, Any] =key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: A__ : str =key.replace("attn" , "attention.self" ) if "fc1" in key: A__ : Dict =key.replace("fc1" , "dense1" ) if "fc2" in key: A__ : str =key.replace("fc2" , "dense2" ) if "linear_pred" in key: A__ : List[Any] =key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: A__ : List[str] =key.replace("linear_fuse.conv" , "linear_fuse" ) A__ : Any =key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 A__ : str =key[key.find("linear_c" ) + len("linear_c" )] A__ : Dict =key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(UpperCamelCase )-1}''' ) if "bot_conv" in key: A__ : Union[str, Any] =key.replace("bot_conv" , "0.convolution" ) if "skip_conv1" in key: A__ : List[Any] =key.replace("skip_conv1" , "1.convolution" ) if "skip_conv2" in key: A__ : int =key.replace("skip_conv2" , "2.convolution" ) if "fusion1" in key: A__ : Optional[Any] =key.replace("fusion1" , "1.fusion" ) if "fusion2" in key: A__ : Optional[Any] =key.replace("fusion2" , "2.fusion" ) if "fusion3" in key: A__ : int =key.replace("fusion3" , "3.fusion" ) if "fusion" in key and "conv" in key: A__ : List[str] =key.replace("conv" , "convolutional_layer" ) if key.startswith("module.last_layer_depth" ): A__ : Tuple =key.replace("module.last_layer_depth" , "head.head" ) A__ : int =value return new_state_dict def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Dict ): """simple docstring""" # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) A__ : int =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) A__ : str =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict A__ : List[str] =kv_weight[ : config.hidden_sizes[i], : ] A__ : Dict =kv_bias[: config.hidden_sizes[i]] A__ : Any =kv_weight[ config.hidden_sizes[i] :, : ] A__ : Any =kv_bias[config.hidden_sizes[i] :] def lowercase ( ): """simple docstring""" A__ : Optional[Any] ="http://images.cocodataset.org/val2017/000000039769.jpg" A__ : List[Any] =Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return image @torch.no_grad() def lowercase ( UpperCamelCase : str , UpperCamelCase : Tuple , UpperCamelCase : List[str]=False , UpperCamelCase : str=None ): """simple docstring""" A__ : List[str] =GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) A__ : str =GLPNImageProcessor() # prepare image A__ : Any =prepare_img() A__ : Optional[int] =image_processor(images=UpperCamelCase , return_tensors="pt" ).pixel_values logger.info("Converting model..." ) # load original state dict A__ : int =torch.load(UpperCamelCase , map_location=torch.device("cpu" ) ) # rename keys A__ : Union[str, Any] =rename_keys(UpperCamelCase ) # key and value matrices need special treatment read_in_k_v(UpperCamelCase , UpperCamelCase ) # create HuggingFace model and load state dict A__ : Optional[int] =GLPNForDepthEstimation(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() # forward pass A__ : int =model(UpperCamelCase ) A__ : Optional[Any] =outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: A__ : List[Any] =torch.tensor( [[4.41_47, 4.08_73, 4.06_73], [3.78_90, 3.28_81, 3.15_25], [3.76_74, 3.54_23, 3.49_13]] ) elif "kitti" in model_name: A__ : Tuple =torch.tensor( [[3.42_91, 2.78_65, 2.51_51], [3.28_41, 2.70_21, 2.35_02], [3.11_47, 2.46_25, 2.24_81]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) A__ : str =torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , UpperCamelCase , atol=1E-4 ) print("Looks ok!" ) # finally, push to hub if required if push_to_hub: logger.info("Pushing model and image processor to the hub..." ) model.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=UpperCamelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=UpperCamelCase , ) if __name__ == "__main__": __A : List[str] = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) parser.add_argument( "--model_name", default="glpn-kitti", type=str, help="Name of the model in case you're pushing to the hub.", ) __A : Any = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
656
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { "caidas/swin2sr-classicalsr-x2-64": ( "https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json" ), } class __magic_name__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = """swin2sr""" _SCREAMING_SNAKE_CASE : Optional[int] = { """hidden_size""": """embed_dim""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : Dict , snake_case_ : str=64 , snake_case_ : Any=1 , snake_case_ : str=3 , snake_case_ : Dict=180 , snake_case_ : List[str]=[6, 6, 6, 6, 6, 6] , snake_case_ : str=[6, 6, 6, 6, 6, 6] , snake_case_ : Any=8 , snake_case_ : Tuple=2.0 , snake_case_ : List[Any]=True , snake_case_ : int=0.0 , snake_case_ : Union[str, Any]=0.0 , snake_case_ : str=0.1 , snake_case_ : Any="gelu" , snake_case_ : int=False , snake_case_ : List[str]=0.02 , snake_case_ : List[str]=1e-5 , snake_case_ : Optional[int]=2 , snake_case_ : Any=1.0 , snake_case_ : Union[str, Any]="1conv" , snake_case_ : Optional[Any]="pixelshuffle" , **snake_case_ : Union[str, Any] , ): super().__init__(**UpperCamelCase__ ) __snake_case = image_size __snake_case = patch_size __snake_case = num_channels __snake_case = embed_dim __snake_case = depths __snake_case = len(UpperCamelCase__ ) __snake_case = num_heads __snake_case = window_size __snake_case = mlp_ratio __snake_case = qkv_bias __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = drop_path_rate __snake_case = hidden_act __snake_case = use_absolute_embeddings __snake_case = layer_norm_eps __snake_case = initializer_range __snake_case = upscale __snake_case = img_range __snake_case = resi_connection __snake_case = upsampler
163
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging __A : Any = logging.get_logger(__name__) __A : Optional[Any] = { "EleutherAI/gpt-neo-1.3B": "https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Union[str, Any] = """gpt_neo""" __magic_name__ : Union[str, Any] = ["""past_key_values"""] __magic_name__ : Dict = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self : Dict , UpperCamelCase__ : List[Any]=50257 , UpperCamelCase__ : Optional[Any]=2048 , UpperCamelCase__ : Tuple=2048 , UpperCamelCase__ : int=24 , UpperCamelCase__ : Dict=[[["global", "local"], 12]] , UpperCamelCase__ : Optional[Any]=16 , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : str=256 , UpperCamelCase__ : List[str]="gelu_new" , UpperCamelCase__ : List[str]=0.0 , UpperCamelCase__ : Tuple=0.0 , UpperCamelCase__ : str=0.0 , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : List[str]=1E-5 , UpperCamelCase__ : Any=0.02 , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : Optional[Any]=50256 , UpperCamelCase__ : List[str]=50256 , **UpperCamelCase__ : str , ): A__ : Optional[Any] =vocab_size A__ : Dict =max_position_embeddings A__ : List[str] =hidden_size A__ : List[Any] =num_layers A__ : Tuple =num_heads A__ : List[str] =intermediate_size A__ : Tuple =window_size A__ : Dict =activation_function A__ : str =resid_dropout A__ : Union[str, Any] =embed_dropout A__ : List[str] =attention_dropout A__ : Tuple =classifier_dropout A__ : int =layer_norm_epsilon A__ : int =initializer_range A__ : str =use_cache A__ : Tuple =bos_token_id A__ : int =eos_token_id A__ : int =attention_types A__ : Any =self.expand_attention_types_params(UpperCamelCase__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.attention_layers)` == `config.num_layers` " F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' "`config.attention_layers` is prepared using `config.attention_types`. " "Please verify the value of `config.attention_types` argument." ) super().__init__(bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) @staticmethod def _UpperCAmelCase ( UpperCamelCase__ : List[str] ): A__ : Optional[Any] =[] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Optional[Any] , UpperCamelCase : List[Any] , UpperCamelCase : Optional[int] ): """simple docstring""" import torch A__ : List[str] =input.size() A__ : Dict =len(UpperCamelCase ) A__ : Optional[int] =shape[dimension] A__ : str =torch.arange(0 , UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =torch.div(sizedim - size , UpperCamelCase , rounding_mode="floor" ) + 1 A__ : str =torch.arange(UpperCamelCase ) + low_indices[:min_length][:, None] A__ : Tuple =[slice(UpperCamelCase )] * rank A__ : int =indices A__ : Optional[int] =input[s] A__ : Union[str, Any] =list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(UpperCamelCase ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any ): """simple docstring""" import torch A__ : List[str] =torch.arange(1 , UpperCamelCase ) A__ : List[Any] =torch.remainder(UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =remainders == 0 A__ : str =candidates[divisor_indices] A__ : int =torch.max(UpperCamelCase ) return largest_divisor, torch.div(UpperCamelCase , UpperCamelCase , rounding_mode="floor" ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' @property def _UpperCAmelCase ( self : List[Any] ): A__ : Optional[int] =OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(UpperCamelCase__ , direction="inputs" ) A__ : Optional[int] ={0: "batch", 1: "past_sequence + sequence"} else: A__ : Tuple ={0: "batch", 1: "sequence"} return common_inputs @property def _UpperCAmelCase ( self : List[str] ): return self._config.num_heads def _UpperCAmelCase ( self : int , UpperCamelCase__ : PreTrainedTokenizer , UpperCamelCase__ : int = -1 , UpperCamelCase__ : int = -1 , UpperCamelCase__ : bool = False , UpperCamelCase__ : Optional[TensorType] = None , ): A__ : Union[str, Any] =super(UpperCamelCase__ , self ).generate_dummy_inputs( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) # We need to order the input in the way they appears in the forward() A__ : List[Any] =OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A__ , A__ : Union[str, Any] =common_inputs["input_ids"].shape # Not using the same length for past_key_values A__ : Union[str, Any] =seqlen + 2 A__ : List[Any] =( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) A__ : Optional[Any] =[ (torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) for _ in range(self.num_layers ) ] A__ : Optional[Any] =common_inputs["attention_mask"] if self.use_past: A__ : Any =ordered_inputs["attention_mask"].dtype A__ : Tuple =torch.cat( [ordered_inputs["attention_mask"], torch.ones(UpperCamelCase__ , UpperCamelCase__ , dtype=UpperCamelCase__ )] , dim=1 ) return ordered_inputs @property def _UpperCAmelCase ( self : List[str] ): return 13
656
0
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer SCREAMING_SNAKE_CASE__ : int =["bert-base-uncased", "bert-base-cased"] SCREAMING_SNAKE_CASE__ : str ="hf-internal-testing/tiny-bert-tf-only" if is_tf_available(): class _UpperCAmelCase ( tf.keras.Model ): """simple docstring""" def __init__( self , _lowercase ) -> Optional[Any]: super().__init__() _lowerCamelCase : str = tokenizer _lowerCamelCase : Any = AutoConfig.from_pretrained(UpperCamelCase__ ) _lowerCamelCase : List[str] = TFAutoModel.from_config(UpperCamelCase__ ) def a__ ( self , _lowercase ) -> List[str]: _lowerCamelCase : List[Any] = self.tokenizer(UpperCamelCase__ ) _lowerCamelCase : Optional[int] = self.bert(**UpperCamelCase__ ) return out["pooler_output"] @require_tf @require_tensorflow_text class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def a__ ( self ) -> Optional[Any]: super().setUp() _lowerCamelCase : int = [ BertTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false _lowerCamelCase : Tuple = [TFBertTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(UpperCamelCase__ , use_fast_bert_tokenizer=UpperCamelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) _lowerCamelCase : str = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] _lowerCamelCase : Optional[Any] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def a__ ( self ) -> Union[str, Any]: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): _lowerCamelCase : List[str] = tokenizer(UpperCamelCase__ , return_tensors='''tf''' , padding='''longest''' ) _lowerCamelCase : int = tf_tokenizer(UpperCamelCase__ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def a__ ( self ) -> str: for tf_tokenizer in self.tf_tokenizers: _lowerCamelCase : Optional[int] = tf_tokenizer(self.paired_sentences ) _lowerCamelCase : Any = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def a__ ( self ) -> List[Any]: for tf_tokenizer in self.tf_tokenizers: _lowerCamelCase : Any = tf.function(UpperCamelCase__ ) for test_inputs in (self.test_sentences, self.paired_sentences): _lowerCamelCase : Optional[int] = tf.constant(UpperCamelCase__ ) _lowerCamelCase : List[Any] = compiled_tokenizer(UpperCamelCase__ ) _lowerCamelCase : str = tf_tokenizer(UpperCamelCase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def a__ ( self ) -> List[Any]: for tf_tokenizer in self.tf_tokenizers: _lowerCamelCase : Dict = ModelToSave(tokenizer=UpperCamelCase__ ) _lowerCamelCase : Optional[Any] = tf.convert_to_tensor(self.test_sentences ) _lowerCamelCase : int = model(UpperCamelCase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: _lowerCamelCase : int = Path(UpperCamelCase__ ) / "saved.model" model.save(UpperCamelCase__ ) _lowerCamelCase : int = tf.keras.models.load_model(UpperCamelCase__ ) _lowerCamelCase : List[Any] = loaded_model(UpperCamelCase__ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
434
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Union[str, Any] = logging.get_logger(__name__) __A : Any = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Tuple = """megatron-bert""" def __init__( self : Tuple , UpperCamelCase__ : Dict=29056 , UpperCamelCase__ : int=1024 , UpperCamelCase__ : Optional[int]=24 , UpperCamelCase__ : Dict=16 , UpperCamelCase__ : int=4096 , UpperCamelCase__ : str="gelu" , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : int=0.1 , UpperCamelCase__ : int=512 , UpperCamelCase__ : str=2 , UpperCamelCase__ : Union[str, Any]=0.02 , UpperCamelCase__ : Any=1E-12 , UpperCamelCase__ : List[Any]=0 , UpperCamelCase__ : str="absolute" , UpperCamelCase__ : Dict=True , **UpperCamelCase__ : Tuple , ): super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) A__ : Optional[int] =vocab_size A__ : Optional[int] =hidden_size A__ : str =num_hidden_layers A__ : Any =num_attention_heads A__ : str =hidden_act A__ : Optional[int] =intermediate_size A__ : str =hidden_dropout_prob A__ : str =attention_probs_dropout_prob A__ : List[Any] =max_position_embeddings A__ : List[Any] =type_vocab_size A__ : Tuple =initializer_range A__ : Any =layer_norm_eps A__ : Any =position_embedding_type A__ : Union[str, Any] =use_cache
656
0
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller __A : Optional[Any] = 3 def __a ( A__ : int ): print("Generating primitive root of p" ) while True: SCREAMING_SNAKE_CASE = random.randrange(3 , A__ ) if pow(A__ , 2 , A__ ) == 1: continue if pow(A__ , A__ , A__ ) == 1: continue return g def __a ( A__ : int ): print("Generating prime p..." ) SCREAMING_SNAKE_CASE = rabin_miller.generate_large_prime(A__ ) # select large prime number. SCREAMING_SNAKE_CASE = primitive_root(A__ ) # one primitive root on modulo p. SCREAMING_SNAKE_CASE = random.randrange(3 , A__ ) # private_key -> have to be greater than 2 for safety. SCREAMING_SNAKE_CASE = cryptomath.find_mod_inverse(pow(A__ , A__ , A__ ) , A__ ) SCREAMING_SNAKE_CASE = (key_size, e_a, e_a, p) SCREAMING_SNAKE_CASE = (key_size, d) return public_key, private_key def __a ( A__ : str , A__ : int ): if os.path.exists(F"{name}_pubkey.txt" ) or os.path.exists(F"{name}_privkey.txt" ): print("\nWARNING:" ) print( F"\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n" "Use a different name or delete these files and re-run this program." ) sys.exit() SCREAMING_SNAKE_CASE = generate_key(A__ ) print(F"\nWriting public key to file {name}_pubkey.txt..." ) with open(F"{name}_pubkey.txt" , "w" ) as fo: fo.write(F"{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}" ) print(F"Writing private key to file {name}_privkey.txt..." ) with open(F"{name}_privkey.txt" , "w" ) as fo: fo.write(F"{private_key[0]},{private_key[1]}" ) def __a ( ): print("Making key files..." ) make_key_files("elgamal" , 2048 ) print("Key files generation successful" ) if __name__ == "__main__": main()
16
"""simple docstring""" from __future__ import annotations def lowercase ( UpperCamelCase : list[float] ): """simple docstring""" if len(UpperCamelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) A__ : Union[str, Any] =nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
656
0
def lowerCamelCase( a__): if length <= 0 or not isinstance(a__ ,a__): raise ValueError('''Length must be a positive integer.''') return [n * (2 * n - 1) for n in range(a__)] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
691
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : Optional[Any] = { "configuration_mega": ["MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegaConfig", "MegaOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ "MEGA_PRETRAINED_MODEL_ARCHIVE_LIST", "MegaForCausalLM", "MegaForMaskedLM", "MegaForMultipleChoice", "MegaForQuestionAnswering", "MegaForSequenceClassification", "MegaForTokenClassification", "MegaModel", "MegaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
from __future__ import annotations _lowerCamelCase : Any = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def a_ ( __lowercase : list[list[int]] , __lowercase : list[int] , __lowercase : list[int] , __lowercase : int , __lowercase : list[list[int]] , ) -> str: _snake_case = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowercase ) ) ] # the reference grid _snake_case = 1 _snake_case = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowercase ) ) ] # the action grid _snake_case = init[0] _snake_case = init[1] _snake_case = 0 _snake_case = g + heuristic[x][y] # cost from starting cell to destination cell _snake_case = [[f, g, x, y]] _snake_case = False # flag that is set when search is complete _snake_case = False # flag set if we can't find expand while not found and not resign: if len(__lowercase ) == 0: raise ValueError('Algorithm is unable to find solution' ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() _snake_case = cell.pop() _snake_case = next_cell[2] _snake_case = next_cell[3] _snake_case = next_cell[1] if x == goal[0] and y == goal[1]: _snake_case = True else: for i in range(len(__lowercase ) ): # to try out different valid actions _snake_case = x + DIRECTIONS[i][0] _snake_case = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__lowercase ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: _snake_case = g + cost _snake_case = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) _snake_case = 1 _snake_case = i _snake_case = [] _snake_case = goal[0] _snake_case = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: _snake_case = x - DIRECTIONS[action[x][y]][0] _snake_case = y - DIRECTIONS[action[x][y]][1] _snake_case = xa _snake_case = ya invpath.append([x, y] ) _snake_case = [] for i in range(len(__lowercase ) ): path.append(invpath[len(__lowercase ) - 1 - i] ) return path, action if __name__ == "__main__": _lowerCamelCase : int = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] _lowerCamelCase : Dict = [0, 0] # all coordinates are given in format [y,x] _lowerCamelCase : Union[str, Any] = [len(grid) - 1, len(grid[0]) - 1] _lowerCamelCase : Any = 1 # the cost map which pushes the path closer to the goal _lowerCamelCase : Dict = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): _lowerCamelCase : Tuple = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map _lowerCamelCase : Union[str, Any] = 99 _lowerCamelCase : Optional[Any] = search(grid, init, goal, cost, heuristic) print('''ACTION MAP''') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
686
"""simple docstring""" def lowercase ( UpperCamelCase : int ): """simple docstring""" if num <= 0: raise ValueError("Input must be a positive integer" ) A__ : Union[str, Any] =[True] * (num + 1) A__ : Union[str, Any] =2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCamelCase ): A__ : str =False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
656
0
"""simple docstring""" from __future__ import annotations import math def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' if depth < 0: raise ValueError("Depth cannot be less than 0" ) if len(_UpperCamelCase ) == 0: raise ValueError("Scores cannot be empty" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , ) return min( minimax(depth + 1 , node_index * 2 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(_UpperCamelCase ) , 2 ) print("Optimal value : " , end="" ) print(minimax(0 , 0 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
636
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : List[Any] ): A__ : Tuple =torch.nn.Linear(10 , 10 ) A__ : List[str] =torch.optim.SGD(model.parameters() , 0.1 ) A__ : Union[str, Any] =Accelerator() A__ : str =accelerator.prepare(UpperCamelCase__ ) try: pickle.loads(pickle.dumps(UpperCamelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowercase = {"configuration_deit": ["DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DeiTConfig", "DeiTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ["DeiTFeatureExtractor"] _lowercase = ["DeiTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ "DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "DeiTForImageClassification", "DeiTForImageClassificationWithTeacher", "DeiTForMaskedImageModeling", "DeiTModel", "DeiTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ "TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDeiTForImageClassification", "TFDeiTForImageClassificationWithTeacher", "TFDeiTForMaskedImageModeling", "TFDeiTModel", "TFDeiTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
91
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __A : Optional[int] = None __A : Union[str, Any] = logging.get_logger(__name__) __A : List[Any] = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __A : str = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } __A : List[str] = { "google/bigbird-roberta-base": 4_096, "google/bigbird-roberta-large": 4_096, "google/bigbird-base-trivia-itc": 4_096, } __A : Tuple = "▁" class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Dict = VOCAB_FILES_NAMES __magic_name__ : Any = PRETRAINED_VOCAB_FILES_MAP __magic_name__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ : List[Any] = BigBirdTokenizer __magic_name__ : Any = ["""input_ids""", """attention_mask"""] __magic_name__ : List[int] = [] def __init__( self : str , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : int="</s>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[Any]="[SEP]" , UpperCamelCase__ : List[Any]="[MASK]" , UpperCamelCase__ : str="[CLS]" , **UpperCamelCase__ : List[Any] , ): A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else bos_token A__ : Optional[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else eos_token A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else unk_token A__ : int =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else pad_token A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else cls_token A__ : List[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) A__ : List[Any] =vocab_file A__ : Optional[int] =False if not self.vocab_file else True def _UpperCAmelCase ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1] def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : Dict =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : List[str] =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
656
0
"""simple docstring""" import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _UpperCamelCase ( A ): UpperCamelCase_ =filter(lambda A : p.requires_grad , model.parameters() ) UpperCamelCase_ =sum([np.prod(p.size() ) for p in model_parameters] ) return params A_ = logging.getLogger(__name__) def _UpperCamelCase ( A , A ): if metric == "rouge2": UpperCamelCase_ ="{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": UpperCamelCase_ ="{val_avg_bleu:.4f}-{step_count}" elif metric == "em": UpperCamelCase_ ="{val_avg_em:.4f}-{step_count}" elif metric == "loss": UpperCamelCase_ ="{val_avg_loss:.4f}-{step_count}" else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" " function." ) UpperCamelCase_ =ModelCheckpoint( dirpath=A , filename=A , monitor=f"""val_{metric}""" , mode="max" , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def _UpperCamelCase ( A , A ): return EarlyStopping( monitor=f"""val_{metric}""" , mode="min" if "loss" in metric else "max" , patience=A , verbose=A , ) class __lowerCAmelCase ( pl.Callback ): '''simple docstring''' def UpperCamelCase__ ( self: Union[str, Any] , UpperCamelCase_: List[Any] , UpperCamelCase_: List[str] ): UpperCamelCase_ ={f"""lr_group_{i}""": param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(UpperCamelCase__ ) @rank_zero_only def UpperCamelCase__ ( self: Optional[int] , UpperCamelCase_: pl.Trainer , UpperCamelCase_: pl.LightningModule , UpperCamelCase_: str , UpperCamelCase_: Any=True ): logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) UpperCamelCase_ =trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results UpperCamelCase_ =Path(pl_module.hparams.output_dir ) if type_path == "test": UpperCamelCase_ =od / "test_results.txt" UpperCamelCase_ =od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. UpperCamelCase_ =od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" UpperCamelCase_ =od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=UpperCamelCase__ ) generations_file.parent.mkdir(exist_ok=UpperCamelCase__ ) with open(UpperCamelCase__ , "a+" ) as writer: for key in sorted(UpperCamelCase__ ): if key in ["log", "progress_bar", "preds"]: continue UpperCamelCase_ =metrics[key] if isinstance(UpperCamelCase__ , torch.Tensor ): UpperCamelCase_ =val.item() UpperCamelCase_ =f"""{key}: {val:.6f}\n""" writer.write(UpperCamelCase__ ) if not save_generations: return if "preds" in metrics: UpperCamelCase_ ="\n".join(metrics["preds"] ) generations_file.open("w+" ).write(UpperCamelCase__ ) @rank_zero_only def UpperCamelCase__ ( self: Optional[int] , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] ): try: UpperCamelCase_ =pl_module.model.model.num_parameters() except AttributeError: UpperCamelCase_ =pl_module.model.num_parameters() UpperCamelCase_ =count_trainable_parameters(UpperCamelCase__ ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1e6, "grad_mp": n_trainable_pars / 1e6} ) @rank_zero_only def UpperCamelCase__ ( self: Any , UpperCamelCase_: pl.Trainer , UpperCamelCase_: pl.LightningModule ): save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(UpperCamelCase__ , UpperCamelCase__ , "test" ) @rank_zero_only def UpperCamelCase__ ( self: int , UpperCamelCase_: pl.Trainer , UpperCamelCase_: str ): save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
391
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Optional[int] = logging.get_logger(__name__) __A : Optional[int] = {"vocab_file": "spiece.model"} __A : List[Any] = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : Dict="<s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : Optional[int]="<sep>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[int]="<cls>" , UpperCamelCase__ : List[str]="<mask>" , UpperCamelCase__ : Optional[Any]=["<eop>", "<eod>"] , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : Dict , ): A__ : List[str] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token A__ : Tuple ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) A__ : Dict =3 A__ : int =do_lower_case A__ : str =remove_space A__ : Optional[Any] =keep_accents A__ : int =vocab_file A__ : Dict =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) A__ : Union[str, Any] =jieba A__ : List[str] =str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _UpperCAmelCase ( self : Union[str, Any] ): return len(self.sp_model ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Any ={self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[str] ): A__ : Union[str, Any] =self.__dict__.copy() A__ : Tuple =None return state def __setstate__( self : Tuple , UpperCamelCase__ : int ): A__ : Union[str, Any] =d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A__ : Optional[int] ={} A__ : Union[str, Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Dict ): if self.remove_space: A__ : Optional[int] =" ".join(inputs.strip().split() ) else: A__ : Optional[Any] =inputs A__ : Any =outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: A__ : Optional[Any] =unicodedata.normalize("NFKD" , UpperCamelCase__ ) A__ : Tuple ="".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: A__ : str =outputs.lower() return outputs def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : str ): A__ : Optional[int] =self.preprocess_text(UpperCamelCase__ ) A__ : Dict =self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) A__ : List[str] =[] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): A__ : str =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : Union[str, Any] =cur_pieces[1:] else: A__ : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def _UpperCAmelCase ( self : int , UpperCamelCase__ : str ): return self.sp_model.PieceToId(UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[Any] ): return self.sp_model.IdToPiece(UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : str ): A__ : Optional[int] ="".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def _UpperCAmelCase ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : Optional[Any] =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : Tuple =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: A__ : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def _UpperCAmelCase ( self : str , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : int ): A__ : List[Any] =super()._decode(*UpperCamelCase__ , **UpperCamelCase__ ) A__ : Union[str, Any] =text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
656
0
'''simple docstring''' import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration __SCREAMING_SNAKE_CASE : Any = pytest.mark.integration __SCREAMING_SNAKE_CASE : List[str] = {"comet"} __SCREAMING_SNAKE_CASE : Dict = importlib.util.find_spec('''fairseq''') is not None __SCREAMING_SNAKE_CASE : List[str] = {"code_eval"} __SCREAMING_SNAKE_CASE : Any = os.name == "nt" __SCREAMING_SNAKE_CASE : Tuple = {"bertscore", "frugalscore", "perplexity"} __SCREAMING_SNAKE_CASE : Optional[int] = importlib.util.find_spec('''transformers''') is not None def a_ ( UpperCamelCase_ ): @wraps(UpperCamelCase_ ) def wrapper(self , UpperCamelCase_ ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("\"test requires Fairseq\"" ) else: test_case(self , UpperCamelCase_ ) return wrapper def a_ ( UpperCamelCase_ ): @wraps(UpperCamelCase_ ) def wrapper(self , UpperCamelCase_ ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("\"test requires transformers\"" ) else: test_case(self , UpperCamelCase_ ) return wrapper def a_ ( UpperCamelCase_ ): @wraps(UpperCamelCase_ ) def wrapper(self , UpperCamelCase_ ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("\"test not supported on Windows\"" ) else: test_case(self , UpperCamelCase_ ) return wrapper def a_ ( ): A_ = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("./metrics/*/" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) @local class __lowerCAmelCase ( parameterized.TestCase ): """simple docstring""" _UpperCAmelCase : int ={} _UpperCAmelCase : Union[str, Any] =None @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning" ) def _UpperCAmelCase ( self : Union[str, Any] , lowerCAmelCase : List[Any] ): A_ = "[...]" A_ = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , UpperCamelCase__ ) ).module_path ) A_ = datasets.load.import_main_class(metric_module.__name__ , dataset=UpperCamelCase__ ) # check parameters A_ = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(UpperCamelCase__ , metric_module.__name__ ): with self.use_local_metrics(): try: A_ = doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def _UpperCAmelCase ( self : str , lowerCAmelCase : Union[str, Any] ): A_ = "[...]" A_ = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , UpperCamelCase__ ) ).module_path ) # run doctest with self.use_local_metrics(): A_ = doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def _UpperCAmelCase ( self : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : Dict ): if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](UpperCamelCase__ ): yield else: yield @contextmanager def _UpperCAmelCase ( self : Union[str, Any] ): def load_local_metric(lowerCAmelCase : Tuple , *lowerCAmelCase : Tuple , **lowerCAmelCase : Tuple ): return load_metric(os.path.join("metrics" , UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ ) with patch("datasets.load_metric" ) as mock_load_metric: A_ = load_local_metric yield @classmethod def _UpperCAmelCase ( cls : Tuple , lowerCAmelCase : str ): def wrapper(lowerCAmelCase : List[Any] ): A_ = contextmanager(UpperCamelCase__ ) A_ = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("bleurt" ) def a_ ( UpperCamelCase_ ): import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("sv" , "" , "" ) # handle pytest cli flags class __lowerCAmelCase ( _UpperCamelCase ): """simple docstring""" def _UpperCAmelCase ( self : Dict , lowerCAmelCase : Optional[Any] ): assert len(input_dict["input_ids"] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("bleurt.score._create_predictor" ) as mock_create_predictor: A_ = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("bertscore" ) def a_ ( UpperCamelCase_ ): import torch def bert_cos_score_idf(UpperCamelCase_ , UpperCamelCase_ , *UpperCamelCase_ , **UpperCamelCase_ ): return torch.tensor([[1.0, 1.0, 1.0]] * len(UpperCamelCase_ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("bert_score.scorer.get_model" ), patch( "bert_score.scorer.bert_cos_score_idf" ) as mock_bert_cos_score_idf: A_ = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("comet" ) def a_ ( UpperCamelCase_ ): def load_from_checkpoint(UpperCamelCase_ ): class __lowerCAmelCase : """simple docstring""" def _UpperCAmelCase ( self : Any , lowerCAmelCase : Dict , *lowerCAmelCase : Tuple , **lowerCAmelCase : List[Any] ): assert len(UpperCamelCase__ ) == 2 A_ = [0.1_9, 0.9_2] return scores, sum(UpperCamelCase__ ) / len(UpperCamelCase__ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("comet.download_model" ) as mock_download_model: A_ = None with patch("comet.load_from_checkpoint" ) as mock_load_from_checkpoint: A_ = load_from_checkpoint yield def a_ ( ): A_ = load_metric(os.path.join("metrics" , "seqeval" ) ) A_ = "ERROR" A_ = f"Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}" with pytest.raises(UpperCamelCase_ , match=re.escape(UpperCamelCase_ ) ): metric.compute(predictions=[] , references=[] , scheme=UpperCamelCase_ )
452
"""simple docstring""" def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations(UpperCamelCase : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations_with_dp_array( UpperCamelCase : int , UpperCamelCase : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] A__ : str =sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase ) for item in array ) A__ : List[str] =answer return answer A__ : List[Any] =[-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase , UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" A__ : str =[0] * (target + 1) A__ : Optional[Any] =1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[Any] = 3 __A : Optional[Any] = 5 __A : int = [1, 2, 5] print(combination_sum_iv(n, array, target))
656
0
"""simple docstring""" from abc import ABC, abstractmethod from typing import List, Optional class __snake_case (_UpperCamelCase ): def __init__( self: List[str] ): # test for the above condition self.test() def __a ( self: int ): __lowerCamelCase = 0 __lowerCamelCase = False while not completed: if counter == 1: self.reset() __lowerCamelCase = self.advance() if not self.does_advance(UpperCamelCase__ ): raise Exception( """Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.""" ) __lowerCamelCase = self.update(UpperCamelCase__ ) counter += 1 if counter > 1_00_00: raise Exception("""update() does not fulfill the constraint.""" ) if self.remaining() != 0: raise Exception("""Custom Constraint is not defined correctly.""" ) @abstractmethod def __a ( self: Tuple ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def __a ( self: Optional[Any] , A_: int ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def __a ( self: Optional[int] , A_: int ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def __a ( self: Any ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def __a ( self: Any ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def __a ( self: List[str] , A_: str=False ): raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) class __snake_case (_UpperCamelCase ): def __init__( self: Optional[int] , A_: List[int] ): super(UpperCamelCase__ , self ).__init__() if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or len(UpperCamelCase__ ) == 0: raise ValueError(f'`token_ids` has to be a non-empty list, but is {token_ids}.' ) if any((not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or token_id < 0) for token_id in token_ids ): raise ValueError(f'Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.' ) __lowerCamelCase = token_ids __lowerCamelCase = len(self.token_ids ) __lowerCamelCase = -1 # the index of the currently fulfilled step __lowerCamelCase = False def __a ( self: Union[str, Any] ): if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def __a ( self: int , A_: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(UpperCamelCase__ )}' ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def __a ( self: Tuple , A_: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(UpperCamelCase__ )}' ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False if self.does_advance(UpperCamelCase__ ): self.fulfilled_idx += 1 __lowerCamelCase = True if self.fulfilled_idx == (self.seqlen - 1): __lowerCamelCase = True __lowerCamelCase = completed else: # failed to make progress. __lowerCamelCase = True self.reset() return stepped, completed, reset def __a ( self: List[str] ): __lowerCamelCase = False __lowerCamelCase = 0 def __a ( self: Tuple ): return self.seqlen - (self.fulfilled_idx + 1) def __a ( self: Optional[int] , A_: Dict=False ): __lowerCamelCase = PhrasalConstraint(self.token_ids ) if stateful: __lowerCamelCase = self.seqlen __lowerCamelCase = self.fulfilled_idx __lowerCamelCase = self.completed return new_constraint class __snake_case : def __init__( self: Optional[int] , A_: List[List[int]] , A_: Optional[int]=True ): __lowerCamelCase = max([len(UpperCamelCase__ ) for one in nested_token_ids] ) __lowerCamelCase = {} for token_ids in nested_token_ids: __lowerCamelCase = root for tidx, token_id in enumerate(UpperCamelCase__ ): if token_id not in level: __lowerCamelCase = {} __lowerCamelCase = level[token_id] if no_subsets and self.has_subsets(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError( """Each list in `nested_token_ids` can't be a complete subset of another list, but is""" f' {nested_token_ids}.' ) __lowerCamelCase = root def __a ( self: int , A_: Optional[int] ): __lowerCamelCase = self.trie for current_token in current_seq: __lowerCamelCase = start[current_token] __lowerCamelCase = list(start.keys() ) return next_tokens def __a ( self: Optional[Any] , A_: Union[str, Any] ): __lowerCamelCase = self.next_tokens(UpperCamelCase__ ) return len(UpperCamelCase__ ) == 0 def __a ( self: Optional[Any] , A_: Optional[Any] ): __lowerCamelCase = list(root.values() ) if len(UpperCamelCase__ ) == 0: return 1 else: return sum([self.count_leaves(UpperCamelCase__ ) for nn in next_nodes] ) def __a ( self: str , A_: Optional[Any] , A_: int ): __lowerCamelCase = self.count_leaves(UpperCamelCase__ ) return len(UpperCamelCase__ ) != leaf_count class __snake_case (_UpperCamelCase ): def __init__( self: List[Any] , A_: List[List[int]] ): super(UpperCamelCase__ , self ).__init__() if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or len(UpperCamelCase__ ) == 0: raise ValueError(f'`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.' ) if any(not isinstance(UpperCamelCase__ , UpperCamelCase__ ) for token_ids in nested_token_ids ): raise ValueError(f'`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.' ) if any( any((not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( f'Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.' ) __lowerCamelCase = DisjunctiveTrie(UpperCamelCase__ ) __lowerCamelCase = nested_token_ids __lowerCamelCase = self.trie.max_height __lowerCamelCase = [] __lowerCamelCase = False def __a ( self: Optional[Any] ): __lowerCamelCase = self.trie.next_tokens(self.current_seq ) if len(UpperCamelCase__ ) == 0: return None else: return token_list def __a ( self: str , A_: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(UpperCamelCase__ )}' ) __lowerCamelCase = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def __a ( self: Tuple , A_: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(UpperCamelCase__ )}' ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False if self.does_advance(UpperCamelCase__ ): self.current_seq.append(UpperCamelCase__ ) __lowerCamelCase = True else: __lowerCamelCase = True self.reset() __lowerCamelCase = self.trie.reached_leaf(self.current_seq ) __lowerCamelCase = completed return stepped, completed, reset def __a ( self: List[str] ): __lowerCamelCase = False __lowerCamelCase = [] def __a ( self: Optional[int] ): if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def __a ( self: Tuple , A_: List[Any]=False ): __lowerCamelCase = DisjunctiveConstraint(self.token_ids ) if stateful: __lowerCamelCase = self.seqlen __lowerCamelCase = self.current_seq __lowerCamelCase = self.completed return new_constraint class __snake_case : def __init__( self: Any , A_: List[Constraint] ): __lowerCamelCase = constraints # max # of steps required to fulfill a given constraint __lowerCamelCase = max([c.seqlen for c in constraints] ) __lowerCamelCase = len(UpperCamelCase__ ) __lowerCamelCase = False self.init_state() def __a ( self: Tuple ): __lowerCamelCase = [] __lowerCamelCase = None __lowerCamelCase = [constraint.copy(stateful=UpperCamelCase__ ) for constraint in self.constraints] def __a ( self: List[Any] ): __lowerCamelCase = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def __a ( self: Any ): __lowerCamelCase = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" __lowerCamelCase = constraint.advance() if isinstance(UpperCamelCase__ , UpperCamelCase__ ): token_list.append(UpperCamelCase__ ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): token_list.extend(UpperCamelCase__ ) else: __lowerCamelCase = self.inprogress_constraint.advance() if isinstance(UpperCamelCase__ , UpperCamelCase__ ): token_list.append(UpperCamelCase__ ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): token_list.extend(UpperCamelCase__ ) if len(UpperCamelCase__ ) == 0: return None else: return token_list def __a ( self: List[Any] , A_: Optional[List[int]] ): self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint __lowerCamelCase = self.add(UpperCamelCase__ ) # the entire list of constraints are fulfilled if self.completed: break def __a ( self: List[Any] , A_: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(f'`token_id` should be an `int`, but is `{token_id}`.' ) __lowerCamelCase = False, False if self.completed: __lowerCamelCase = True __lowerCamelCase = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state __lowerCamelCase = self.inprogress_constraint.update(UpperCamelCase__ ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=UpperCamelCase__ ) ) __lowerCamelCase = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) __lowerCamelCase = None if len(self.pending_constraints ) == 0: # we're done! __lowerCamelCase = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(UpperCamelCase__ ): __lowerCamelCase = pending_constraint.update(UpperCamelCase__ ) if not stepped: raise Exception( """`constraint.update(token_id)` is not yielding incremental progress, """ """even though `constraint.does_advance(token_id)` is true.""" ) if complete: self.complete_constraints.append(UpperCamelCase__ ) __lowerCamelCase = None if not complete and stepped: __lowerCamelCase = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". __lowerCamelCase = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. __lowerCamelCase = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def __a ( self: Dict , A_: Optional[int]=True ): __lowerCamelCase = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: __lowerCamelCase = [ constraint.copy(stateful=UpperCamelCase__ ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: __lowerCamelCase = self.inprogress_constraint.copy(stateful=UpperCamelCase__ ) __lowerCamelCase = [constraint.copy() for constraint in self.pending_constraints] return new_state
281
"""simple docstring""" import math import tensorflow as tf from packaging import version def lowercase ( UpperCamelCase : Optional[Any] ): """simple docstring""" A__ : List[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : Optional[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : Tuple =tf.cast(math.pi , x.dtype ) A__ : Dict =tf.cast(0.04_47_15 , x.dtype ) A__ : Optional[int] =0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(UpperCamelCase , 3 )) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) return x * tf.tanh(tf.math.softplus(UpperCamelCase ) ) def lowercase ( UpperCamelCase : List[str] ): """simple docstring""" A__ : Union[str, Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =tf.cast(0.04_47_15 , x.dtype ) A__ : List[Any] =tf.cast(0.79_78_84_56_08 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) A__ : str =tf.cast(1.7_02 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" return tf.clip_by_value(_gelu(UpperCamelCase ) , -10 , 10 ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any=-1 ): """simple docstring""" A__ , A__ : Optional[Any] =tf.split(UpperCamelCase , 2 , axis=UpperCamelCase ) return a * tf.math.sigmoid(UpperCamelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def lowercase ( UpperCamelCase : int ): """simple docstring""" return tf.keras.activations.gelu(UpperCamelCase , approximate=UpperCamelCase ) __A : Optional[Any] = tf.keras.activations.gelu __A : Optional[Any] = approximate_gelu_wrap else: __A : Any = _gelu __A : Union[str, Any] = _gelu_new __A : List[str] = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F'''function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}''' )
656
0
def a__ (__lowercase :Tuple , __lowercase :List[str] ) -> Tuple: _A : Optional[Any] = [1] for i in range(2 , __lowercase ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" _A : Optional[Any] = [] _A : Union[str, Any] = list(range(__lowercase ) ) # Find permutation while factorials: _A : Optional[int] = factorials.pop() _A : Optional[Any] = divmod(__lowercase , __lowercase ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
206
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict ): A__ : Optional[Any] =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_attention_heads" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_encoder_blocks" ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any]=13 , UpperCamelCase__ : Tuple=64 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Union[str, Any]=4 , UpperCamelCase__ : Dict=[2, 2, 2, 2] , UpperCamelCase__ : Union[str, Any]=[8, 4, 2, 1] , UpperCamelCase__ : Tuple=[16, 32, 64, 128] , UpperCamelCase__ : Optional[int]=[1, 4, 8, 16] , UpperCamelCase__ : Any=[1, 2, 4, 8] , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Dict="gelu" , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[Any]=None , ): A__ : Tuple =parent A__ : List[Any] =batch_size A__ : List[Any] =image_size A__ : Union[str, Any] =num_channels A__ : Optional[int] =num_encoder_blocks A__ : Any =sr_ratios A__ : Any =depths A__ : List[Any] =hidden_sizes A__ : List[Any] =downsampling_rates A__ : List[str] =num_attention_heads A__ : int =is_training A__ : List[Any] =use_labels A__ : Any =hidden_act A__ : Dict =hidden_dropout_prob A__ : int =attention_probs_dropout_prob A__ : List[Any] =initializer_range A__ : Tuple =num_labels A__ : List[Any] =scope def _UpperCAmelCase ( self : Optional[int] ): A__ : List[str] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Any =None if self.use_labels: A__ : Tuple =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A__ : List[Any] =self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self : Tuple ): return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int ): A__ : Any =SegformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Dict =model(UpperCamelCase__ ) A__ : Optional[int] =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[Any] ): A__ : str =self.num_labels A__ : Optional[Any] =SegformerForSemanticSegmentation(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Optional[Any] =model(UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) A__ : List[Any] =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str ): A__ : Tuple =1 A__ : Tuple =SegformerForSemanticSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : List[str] =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(UpperCamelCase__ ) A__ : Dict =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : str ): A__ : Union[str, Any] =self.prepare_config_and_inputs() A__ , A__ , A__ : Tuple =config_and_inputs A__ : Tuple ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Dict = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __magic_name__ : Optional[int] = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ : Dict = True __magic_name__ : List[str] = False __magic_name__ : Optional[Any] = False __magic_name__ : str = False def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =SegformerModelTester(self ) A__ : Tuple =SegformerConfigTester(self , config_class=UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): self.config_tester.run_common_tests() def _UpperCAmelCase ( self : Dict ): A__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*UpperCamelCase__ ) @unittest.skip("SegFormer does not use inputs_embeds" ) def _UpperCAmelCase ( self : Dict ): pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods" ) def _UpperCAmelCase ( self : Tuple ): pass def _UpperCAmelCase ( self : List[str] ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : int =model_class(UpperCamelCase__ ) A__ : Optional[int] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Optional[int] =[*signature.parameters.keys()] A__ : List[str] =["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() A__ : Union[str, Any] =True for model_class in self.all_model_classes: A__ : Optional[Any] =True A__ : Union[str, Any] =False A__ : str =True A__ : Optional[int] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : str =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Any =outputs.attentions A__ : List[str] =sum(self.model_tester.depths ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ : Dict =True A__ : str =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Any =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Union[str, Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : List[Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) A__ : Tuple =(self.model_tester.image_size // 32) ** 2 A__ : Optional[Any] =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) A__ : int =len(UpperCamelCase__ ) # Check attention is always last and order is fine A__ : Optional[Any] =True A__ : Any =True A__ : Union[str, Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Optional[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(out_len + 1 , len(UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : Union[str, Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _UpperCAmelCase ( self : List[Any] ): def check_hidden_states_output(UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple ): A__ : Optional[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : List[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.hidden_states A__ : int =self.model_tester.num_encoder_blocks self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Optional[Any] =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : str =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[int] ): if not self.model_tester.is_training: return A__ , A__ : int =self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] =True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ): continue A__ : List[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() A__ : int =self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) A__ : Union[str, Any] =model(**UpperCamelCase__ ).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _UpperCAmelCase ( self : Tuple ): pass @slow def _UpperCAmelCase ( self : Tuple ): for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Tuple =SegformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowercase ( ): """simple docstring""" A__ : List[Any] =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): # only resize + normalize A__ : List[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : Union[str, Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : Union[str, Any] =prepare_img() A__ : Union[str, Any] =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : int =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : Dict =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : Optional[int] =torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : Union[str, Any] ): # only resize + normalize A__ : Dict =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : int =SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024" ).to(UpperCamelCase__ ) A__ : Tuple =prepare_img() A__ : str =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Optional[int] =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : List[str] =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : List[Any] =torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-1 ) ) @slow def _UpperCAmelCase ( self : int ): # only resize + normalize A__ : Optional[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : List[Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : str =prepare_img() A__ : Dict =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Any =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : Dict =model(UpperCamelCase__ ) A__ : Any =outputs.logits.detach().cpu() A__ : Union[str, Any] =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ , target_sizes=[(500, 300)] ) A__ : List[str] =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ ) A__ : int =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ ) A__ : Tuple =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ )
656
0
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __UpperCamelCase ( SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" __snake_case = [2, 2, 6, 2] if "tiny" in model_name else [2, 2, 18, 2] __snake_case = True if "large" in model_name or "huge" in model_name else False __snake_case = True if "large" in model_name or "huge" in model_name else False __snake_case = True if "large" in model_name or "huge" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: __snake_case = [3, 3, 3, 3] __snake_case = [5, 5, 5, 5] elif "fl4" in model_name: __snake_case = [4, 4, 4, 4] __snake_case = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: __snake_case = [3, 3, 3, 3] if "lrf" in model_name: __snake_case = [3, 3, 3, 3] else: __snake_case = [2, 2, 2, 2] if "tiny" in model_name: __snake_case = 96 elif "small" in model_name: __snake_case = 96 elif "base" in model_name: __snake_case = 1_28 elif "large" in model_name: __snake_case = 1_92 elif "xlarge" in model_name: __snake_case = 2_56 elif "huge" in model_name: __snake_case = 3_52 # set label information __snake_case = "huggingface/label-files" if "large" in model_name or "huge" in model_name: __snake_case = "imagenet-22k-id2label.json" else: __snake_case = "imagenet-1k-id2label.json" __snake_case = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type="dataset" ) , "r" ) ) __snake_case = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} __snake_case = {v: k for k, v in idalabel.items()} __snake_case = FocalNetConfig( embed_dim=SCREAMING_SNAKE_CASE , depths=SCREAMING_SNAKE_CASE , focal_levels=SCREAMING_SNAKE_CASE , focal_windows=SCREAMING_SNAKE_CASE , use_conv_embed=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , use_post_layernorm=SCREAMING_SNAKE_CASE , use_layerscale=SCREAMING_SNAKE_CASE , ) return config def __UpperCamelCase ( SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" if "patch_embed.proj" in name: __snake_case = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: __snake_case = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: __snake_case = "encoder." + name if "encoder.layers" in name: __snake_case = name.replace("encoder.layers" , "encoder.stages" ) if "downsample.proj" in name: __snake_case = name.replace("downsample.proj" , "downsample.projection" ) if "blocks" in name: __snake_case = name.replace("blocks" , "layers" ) if "modulation.f.weight" in name or "modulation.f.bias" in name: __snake_case = name.replace("modulation.f" , "modulation.projection_in" ) if "modulation.h.weight" in name or "modulation.h.bias" in name: __snake_case = name.replace("modulation.h" , "modulation.projection_context" ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: __snake_case = name.replace("modulation.proj" , "modulation.projection_out" ) if name == "norm.weight": __snake_case = "layernorm.weight" if name == "norm.bias": __snake_case = "layernorm.bias" if "head" in name: __snake_case = name.replace("head" , "classifier" ) else: __snake_case = "focalnet." + name return name def __UpperCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ) -> Dict: """simple docstring""" __snake_case = { "focalnet-tiny": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth", "focalnet-tiny-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth", "focalnet-small": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth", "focalnet-small-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth", "focalnet-base": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth", "focalnet-base-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth", "focalnet-large-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth", "focalnet-large-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth", "focalnet-xlarge-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth", "focalnet-xlarge-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth", } # fmt: on __snake_case = model_name_to_url[model_name] print("Checkpoint URL: " , SCREAMING_SNAKE_CASE ) __snake_case = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location="cpu" )["model"] # rename keys for key in state_dict.copy().keys(): __snake_case = state_dict.pop(SCREAMING_SNAKE_CASE ) __snake_case = val __snake_case = get_focalnet_config(SCREAMING_SNAKE_CASE ) __snake_case = FocalNetForImageClassification(SCREAMING_SNAKE_CASE ) model.eval() # load state dict model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify conversion __snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE , size={"shortest_edge": 2_56} , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE , crop_size=2_24 , do_normalize=SCREAMING_SNAKE_CASE , image_mean=SCREAMING_SNAKE_CASE , image_std=SCREAMING_SNAKE_CASE , ) __snake_case = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ) __snake_case = processor(images=SCREAMING_SNAKE_CASE , return_tensors="pt" ) __snake_case = transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) __snake_case = image_transforms(SCREAMING_SNAKE_CASE ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , SCREAMING_SNAKE_CASE , atol=1e-4 ) __snake_case = model(**SCREAMING_SNAKE_CASE ) __snake_case = outputs.logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) print("First values of logits:" , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": __snake_case = torch.tensor([0.2_166, -0.4_368, 0.2_191] ) elif model_name == "focalnet-tiny-lrf": __snake_case = torch.tensor([1.1_669, 0.0_125, -0.1_695] ) elif model_name == "focalnet-small": __snake_case = torch.tensor([0.4_917, -0.0_430, 0.1_341] ) elif model_name == "focalnet-small-lrf": __snake_case = torch.tensor([-0.2_588, -0.5_342, -0.2_331] ) elif model_name == "focalnet-base": __snake_case = torch.tensor([-0.1_655, -0.4_090, -0.1_730] ) elif model_name == "focalnet-base-lrf": __snake_case = torch.tensor([0.5_306, -0.0_483, -0.3_928] ) assert torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""focalnet-tiny""", type=str, help="""Name of the FocalNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub.""", ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
163
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any]=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[str]=99 , UpperCamelCase__ : Optional[Any]=32 , UpperCamelCase__ : Any=5 , UpperCamelCase__ : Dict=4 , UpperCamelCase__ : Union[str, Any]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Any=0.1 , UpperCamelCase__ : Optional[Any]=0.1 , UpperCamelCase__ : Optional[Any]=512 , UpperCamelCase__ : Union[str, Any]=16 , UpperCamelCase__ : List[str]=2 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : List[Any]=4 , ): A__ : str =parent A__ : List[str] =batch_size A__ : Any =seq_length A__ : List[str] =is_training A__ : List[Any] =use_attention_mask A__ : List[Any] =use_token_type_ids A__ : Dict =use_labels A__ : List[Any] =vocab_size A__ : Optional[int] =hidden_size A__ : Optional[Any] =num_hidden_layers A__ : str =num_attention_heads A__ : int =intermediate_size A__ : Tuple =hidden_act A__ : Tuple =hidden_dropout_prob A__ : Dict =attention_probs_dropout_prob A__ : Any =max_position_embeddings A__ : Any =type_vocab_size A__ : Union[str, Any] =type_sequence_label_size A__ : Optional[Any] =initializer_range A__ : int =num_choices def _UpperCAmelCase ( self : Tuple ): A__ : List[str] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[str] =None if self.use_attention_mask: A__ : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) A__ : str =None if self.use_token_type_ids: A__ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Any =RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCAmelCase ( self : Tuple ): A__ : Dict =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : str =config_and_inputs A__ : Optional[Any] ={"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _UpperCAmelCase ( self : int ): A__ : str =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Union[str, Any] =config_and_inputs A__ : Union[str, Any] =True A__ : List[Any] =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : Tuple =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __lowerCAmelCase ( _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Union[str, Any] = True __magic_name__ : Dict = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Optional[int] =FlaxRobertaPreLayerNormModelTester(self ) @slow def _UpperCAmelCase ( self : List[Any] ): for model_class_name in self.all_model_classes: A__ : Tuple =model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Union[str, Any] =model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): A__ : Any =FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Tuple =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : str =model(UpperCamelCase__ )[0] A__ : List[Any] =[1, 11, 50265] self.assertEqual(list(output.shape ) , UpperCamelCase__ ) # compare the actual values for a slice. A__ : Any =np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : List[Any] ): A__ : Union[str, Any] =FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : List[Any] =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : Dict =model(UpperCamelCase__ )[0] # compare the actual values for a slice. A__ : Optional[Any] =np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) )
656
0
"""simple docstring""" import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def a__ ( self ) -> str: _lowerCamelCase : List[Any] = tempfile.mkdtemp() _lowerCamelCase : Dict = 8 # DPR tok _lowerCamelCase : Union[str, Any] = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) _lowerCamelCase : Dict = os.path.join(UpperCamelCase__ , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok _lowerCamelCase : Any = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] _lowerCamelCase : Union[str, Any] = dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) _lowerCamelCase : str = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] _lowerCamelCase : Tuple = {"unk_token": "<unk>"} _lowerCamelCase : Any = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) _lowerCamelCase : List[str] = os.path.join(UpperCamelCase__ , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase : Dict = os.path.join(UpperCamelCase__ , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCamelCase__ ) ) def a__ ( self ) -> Tuple: return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def a__ ( self ) -> str: return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def a__ ( self ) -> List[str]: return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def a__ ( self ) -> int: shutil.rmtree(self.tmpdirname ) def a__ ( self ) -> Dict: _lowerCamelCase : Any = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def a__ ( self ) -> Union[str, Any]: _lowerCamelCase : List[str] = self.get_dummy_dataset() _lowerCamelCase : Optional[Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: _lowerCamelCase : List[Any] = dataset _lowerCamelCase : Tuple = RagRetriever( UpperCamelCase__ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def a__ ( self , _lowercase ) -> Optional[Any]: _lowerCamelCase : List[Any] = self.get_dummy_dataset() _lowerCamelCase : int = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: _lowerCamelCase : Optional[Any] = os.path.join(self.tmpdirname , '''dataset''' ) _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset _lowerCamelCase : Union[str, Any] = RagRetriever( UpperCamelCase__ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: _lowerCamelCase : Union[str, Any] = RagRetriever( UpperCamelCase__ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , UpperCamelCase__ ) , ) return retriever def a__ ( self ) -> List[str]: _lowerCamelCase : Tuple = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) _lowerCamelCase : str = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) _lowerCamelCase : List[Any] = {sample["id"]: [sample["text"], sample["title"]] for sample in dataset} pickle.dump(UpperCamelCase__ , open(UpperCamelCase__ , '''wb''' ) ) _lowerCamelCase : str = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) _lowerCamelCase : List[str] = RagRetriever( UpperCamelCase__ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def a__ ( self ) -> Dict: _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : str = self.get_dummy_canonical_hf_index_retriever() _lowerCamelCase : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Optional[int] = retriever.retrieve(UpperCamelCase__ , n_docs=UpperCamelCase__ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCamelCase__ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCamelCase__ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def a__ ( self ) -> List[Any]: _lowerCamelCase : int = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: _lowerCamelCase : Dict = self.get_dummy_dataset() retriever.save_pretrained(UpperCamelCase__ ) _lowerCamelCase : List[Any] = RagRetriever.from_pretrained(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) _lowerCamelCase : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : List[str] = retriever.retrieve(UpperCamelCase__ , n_docs=1 ) self.assertTrue(out is not None ) def a__ ( self ) -> Union[str, Any]: _lowerCamelCase : Dict = 1 _lowerCamelCase : Tuple = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCamelCase__ ) _lowerCamelCase : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Any = retriever.retrieve(UpperCamelCase__ , n_docs=UpperCamelCase__ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCamelCase__ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCamelCase__ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def a__ ( self ) -> Any: _lowerCamelCase : Optional[Any] = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCamelCase__ ) _lowerCamelCase : List[str] = RagRetriever.from_pretrained(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) _lowerCamelCase : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Union[str, Any] = retriever.retrieve(UpperCamelCase__ , n_docs=1 ) self.assertTrue(out is not None ) def a__ ( self ) -> Tuple: _lowerCamelCase : List[str] = 1 _lowerCamelCase : str = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCamelCase__ ) _lowerCamelCase : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Any = retriever.retrieve(UpperCamelCase__ , n_docs=UpperCamelCase__ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCamelCase__ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , UpperCamelCase__ ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def a__ ( self ) -> Dict: _lowerCamelCase : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCamelCase__ ) _lowerCamelCase : Any = RagRetriever.from_pretrained(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) _lowerCamelCase : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : str = retriever.retrieve(UpperCamelCase__ , n_docs=1 ) self.assertTrue(out is not None ) def a__ ( self ) -> Dict: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[str] = self.get_dummy_legacy_index_retriever() _lowerCamelCase : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Any = retriever.retrieve(UpperCamelCase__ , n_docs=UpperCamelCase__ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(UpperCamelCase__ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , UpperCamelCase__ ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def a__ ( self ) -> Dict: _lowerCamelCase : List[str] = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(UpperCamelCase__ ) _lowerCamelCase : Optional[Any] = RagRetriever.from_pretrained(UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) _lowerCamelCase : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Dict = retriever.retrieve(UpperCamelCase__ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def a__ ( self ) -> Any: import torch _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : List[str] = self.get_dummy_canonical_hf_index_retriever() _lowerCamelCase : str = [[5, 7], [10, 11]] _lowerCamelCase : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Optional[int] = retriever(UpperCamelCase__ , UpperCamelCase__ , prefix=retriever.config.generator.prefix , n_docs=UpperCamelCase__ ) _lowerCamelCase : str = ( out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , UpperCamelCase__ ) self.assertIsInstance(UpperCamelCase__ , np.ndarray ) _lowerCamelCase : Dict = retriever( UpperCamelCase__ , UpperCamelCase__ , prefix=retriever.config.generator.prefix , n_docs=UpperCamelCase__ , return_tensors='''pt''' , ) _lowerCamelCase : List[str] = ( # noqa: F841 out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], out["doc_ids"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) self.assertIsInstance(UpperCamelCase__ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def a__ ( self ) -> str: _lowerCamelCase : List[Any] = self.get_dpr_ctx_encoder_tokenizer() _lowerCamelCase : int = 1 _lowerCamelCase : Any = self.get_dummy_custom_hf_index_retriever(from_disk=UpperCamelCase__ ) retriever.set_ctx_encoder_tokenizer(UpperCamelCase__ ) _lowerCamelCase : List[Any] = [[5, 7], [10, 11]] _lowerCamelCase : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) _lowerCamelCase : Tuple = retriever(UpperCamelCase__ , UpperCamelCase__ , prefix=retriever.config.generator.prefix , n_docs=UpperCamelCase__ ) self.assertEqual( len(UpperCamelCase__ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , UpperCamelCase__ ) # check for doc token related keys in dictionary.
434
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __A : List[Any] = logging.get_logger(__name__) __A : Any = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] __A : Optional[int] = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" A__ : Union[str, Any] =torch.load(UpperCamelCase , map_location="cpu" ) return sd def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : int=rename_keys_prefix ): """simple docstring""" A__ : List[str] =OrderedDict() A__ : str =torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue A__ : Optional[Any] =key for name_pair in rename_keys_prefix: A__ : int =new_key.replace(name_pair[0] , name_pair[1] ) A__ : Dict =d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately A__ : Optional[int] =new_d["cls.predictions.bias"] return new_d @torch.no_grad() def lowercase ( UpperCamelCase : Dict , UpperCamelCase : List[str] ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: A__ : Any ="pretraining" if "vcr" in checkpoint_path: A__ : Union[str, Any] ={"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: A__ : Optional[Any] ={"visual_embedding_dim": 2048} elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048} elif "nlvr" in checkpoint_path: A__ : List[str] ={"visual_embedding_dim": 1024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 512} A__ : List[str] ="multichoice" elif "vqa_advanced" in checkpoint_path: A__ : Any ={"visual_embedding_dim": 2048} A__ : str ="vqa_advanced" elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048, "num_labels": 3129} A__ : str ="vqa" elif "nlvr" in checkpoint_path: A__ : str ={ "visual_embedding_dim": 1024, "num_labels": 2, } A__ : Dict ="nlvr" A__ : Union[str, Any] =VisualBertConfig(**UpperCamelCase ) # Load State Dict A__ : int =load_state_dict(UpperCamelCase ) A__ : Tuple =get_new_dict(UpperCamelCase , UpperCamelCase ) if model_type == "pretraining": A__ : str =VisualBertForPreTraining(UpperCamelCase ) elif model_type == "vqa": A__ : Optional[int] =VisualBertForQuestionAnswering(UpperCamelCase ) elif model_type == "nlvr": A__ : Union[str, Any] =VisualBertForVisualReasoning(UpperCamelCase ) elif model_type == "multichoice": A__ : Union[str, Any] =VisualBertForMultipleChoice(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) # Save Checkpoints Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": __A : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("orig_checkpoint_path", type=str, help="A path to .th on local filesystem.") parser.add_argument("pytorch_dump_folder_path", type=str, help="Path to the output PyTorch model.") __A : str = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
656
0
from __future__ import annotations def __a ( A__ : str ): return [ord(A__ ) - 96 for elem in plain] def __a ( A__ : list[int] ): return "".join(chr(elem + 96 ) for elem in encoded ) def __a ( ): SCREAMING_SNAKE_CASE = encode(input("-> " ).strip().lower() ) print("Encoded: " , A__ ) print("Decoded:" , decode(A__ ) ) if __name__ == "__main__": main()
16
"""simple docstring""" __A : Union[str, Any] = {str(digit): digit**5 for digit in range(10)} def lowercase ( UpperCamelCase : int ): """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase ) ) def lowercase ( ): """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(UpperCamelCase ) ) if __name__ == "__main__": print(solution())
656
0
import argparse import math import os import torch from neural_compressor.utils.pytorch import load from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, StableDiffusionPipeline, UNetaDConditionModel def lowerCamelCase( ): _SCREAMING_SNAKE_CASE =argparse.ArgumentParser() parser.add_argument( '''-m''' ,'''--pretrained_model_name_or_path''' ,type=a__ ,default=a__ ,required=a__ ,help='''Path to pretrained model or model identifier from huggingface.co/models.''' ,) parser.add_argument( '''-c''' ,'''--caption''' ,type=a__ ,default='''robotic cat with wings''' ,help='''Text used to generate images.''' ,) parser.add_argument( '''-n''' ,'''--images_num''' ,type=a__ ,default=4 ,help='''How much images to generate.''' ,) parser.add_argument( '''-s''' ,'''--seed''' ,type=a__ ,default=42 ,help='''Seed for random process.''' ,) parser.add_argument( '''-ci''' ,'''--cuda_id''' ,type=a__ ,default=0 ,help='''cuda_id.''' ,) _SCREAMING_SNAKE_CASE =parser.parse_args() return args def lowerCamelCase( a__ ,a__ ,a__): if not len(a__) == rows * cols: raise ValueError('''The specified number of rows and columns are not correct.''') _SCREAMING_SNAKE_CASE =imgs[0].size _SCREAMING_SNAKE_CASE =Image.new('''RGB''' ,size=(cols * w, rows * h)) _SCREAMING_SNAKE_CASE =grid.size for i, img in enumerate(a__): grid.paste(a__ ,box=(i % cols * w, i // cols * h)) return grid def lowerCamelCase( a__ ,a__="robotic cat with wings" ,a__=7.5 ,a__=50 ,a__=1 ,a__=42 ,): _SCREAMING_SNAKE_CASE =torch.Generator(pipeline.device).manual_seed(a__) _SCREAMING_SNAKE_CASE =pipeline( a__ ,guidance_scale=a__ ,num_inference_steps=a__ ,generator=a__ ,num_images_per_prompt=a__ ,).images _SCREAMING_SNAKE_CASE =int(math.sqrt(a__)) _SCREAMING_SNAKE_CASE =image_grid(a__ ,rows=_rows ,cols=num_images_per_prompt // _rows) return grid, images snake_case_ : Dict = parse_args() # Load models and create wrapper for stable diffusion snake_case_ : List[str] = CLIPTokenizer.from_pretrained(args.pretrained_model_name_or_path, subfolder='''tokenizer''') snake_case_ : List[str] = CLIPTextModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='''text_encoder''') snake_case_ : List[str] = AutoencoderKL.from_pretrained(args.pretrained_model_name_or_path, subfolder='''vae''') snake_case_ : Dict = UNetaDConditionModel.from_pretrained(args.pretrained_model_name_or_path, subfolder='''unet''') snake_case_ : Tuple = StableDiffusionPipeline.from_pretrained( args.pretrained_model_name_or_path, text_encoder=text_encoder, vae=vae, unet=unet, tokenizer=tokenizer ) snake_case_ : List[str] = lambda images, clip_input: (images, False) if os.path.exists(os.path.join(args.pretrained_model_name_or_path, '''best_model.pt''')): snake_case_ : Dict = load(args.pretrained_model_name_or_path, model=unet) unet.eval() setattr(pipeline, '''unet''', unet) else: snake_case_ : int = unet.to(torch.device('''cuda''', args.cuda_id)) snake_case_ : Dict = pipeline.to(unet.device) snake_case_ : Optional[int] = generate_images(pipeline, prompt=args.caption, num_images_per_prompt=args.images_num, seed=args.seed) grid.save(os.path.join(args.pretrained_model_name_or_path, '''{}.png'''.format('''_'''.join(args.caption.split())))) snake_case_ : Optional[int] = os.path.join(args.pretrained_model_name_or_path, '''_'''.join(args.caption.split())) os.makedirs(dirname, exist_ok=True) for idx, image in enumerate(images): image.save(os.path.join(dirname, '''{}.png'''.format(idx + 1)))
691
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig __A : Optional[Any] = logging.get_logger(__name__) # General docstring __A : str = "PoolFormerConfig" # Base docstring __A : Optional[Any] = "sail/poolformer_s12" __A : List[Any] = [1, 512, 7, 7] # Image classification docstring __A : List[str] = "sail/poolformer_s12" __A : Tuple = "tabby, tabby cat" __A : Tuple = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def lowercase ( UpperCamelCase : Any , UpperCamelCase : float = 0.0 , UpperCamelCase : bool = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input A__ : Tuple =1 - drop_prob A__ : List[str] =(input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets A__ : Any =keep_prob + torch.rand(UpperCamelCase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize A__ : Optional[int] =input.div(UpperCamelCase ) * random_tensor return output class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : Optional[float] = None ): super().__init__() A__ : Optional[int] =drop_prob def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : torch.Tensor ): return drop_path(UpperCamelCase__ , self.drop_prob , self.training ) def _UpperCAmelCase ( self : List[str] ): return "p={}".format(self.drop_prob ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): super().__init__() A__ : Optional[int] =patch_size if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (patch_size, patch_size) A__ : Optional[int] =stride if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (stride, stride) A__ : int =padding if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (padding, padding) A__ : Any =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=UpperCamelCase__ ) A__ : Any =norm_layer(UpperCamelCase__ ) if norm_layer else nn.Identity() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : str ): A__ : List[str] =self.projection(UpperCamelCase__ ) A__ : Any =self.norm(UpperCamelCase__ ) return embeddings class __lowerCAmelCase ( nn.GroupNorm): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Dict , **UpperCamelCase__ : Union[str, Any] ): super().__init__(1 , UpperCamelCase__ , **UpperCamelCase__ ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Optional[int] ): super().__init__() A__ : Any =nn.AvgPoolad(UpperCamelCase__ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[str] ): return self.pool(UpperCamelCase__ ) - hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] ): super().__init__() A__ : List[Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Union[str, Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Dict =PoolFormerDropPath(UpperCamelCase__ ) if isinstance(config.hidden_act , UpperCamelCase__ ): A__ : Tuple =ACTaFN[config.hidden_act] else: A__ : Optional[Any] =config.hidden_act def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict ): A__ : Optional[Any] =self.conva(UpperCamelCase__ ) A__ : List[str] =self.act_fn(UpperCamelCase__ ) A__ : List[str] =self.drop(UpperCamelCase__ ) A__ : Optional[int] =self.conva(UpperCamelCase__ ) A__ : Optional[Any] =self.drop(UpperCamelCase__ ) return hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Any ): super().__init__() A__ : Optional[int] =PoolFormerPooling(UpperCamelCase__ ) A__ : List[str] =PoolFormerOutput(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) # Useful for training neural nets A__ : Tuple =PoolFormerDropPath(UpperCamelCase__ ) if drop_path > 0.0 else nn.Identity() A__ : Optional[Any] =config.use_layer_scale if config.use_layer_scale: A__ : List[str] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) A__ : List[Any] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : Optional[int] ): if self.use_layer_scale: A__ : Optional[int] =self.pooling(self.before_norm(UpperCamelCase__ ) ) A__ : Union[str, Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection A__ : Union[str, Any] =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : Tuple =() A__ : List[str] =self.output(self.after_norm(UpperCamelCase__ ) ) A__ : Optional[Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection A__ : str =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : List[Any] =(output,) + outputs return outputs else: A__ : Tuple =self.drop_path(self.pooling(self.before_norm(UpperCamelCase__ ) ) ) # First residual connection A__ : Optional[Any] =pooling_output + hidden_states A__ : Tuple =() # Second residual connection inside the PoolFormerOutput block A__ : List[str] =self.drop_path(self.output(self.after_norm(UpperCamelCase__ ) ) ) A__ : Any =hidden_states + layer_output A__ : Tuple =(output,) + outputs return outputs class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : List[str] ): super().__init__() A__ : Tuple =config # stochastic depth decay rule A__ : Dict =[x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings A__ : Tuple =[] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) A__ : List[str] =nn.ModuleList(UpperCamelCase__ ) # Transformer blocks A__ : Union[str, Any] =[] A__ : Any =0 for i in range(config.num_encoder_blocks ): # each block consists of layers A__ : Union[str, Any] =[] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCamelCase__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCamelCase__ ) ) A__ : str =nn.ModuleList(UpperCamelCase__ ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : Optional[int]=True ): A__ : Union[str, Any] =() if output_hidden_states else None A__ : Dict =pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): A__ , A__ : List[Any] =layers # Get patch embeddings from hidden_states A__ : Any =embedding_layer(UpperCamelCase__ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCamelCase__ ): A__ : List[str] =blk(UpperCamelCase__ ) A__ : Tuple =layer_outputs[0] if output_hidden_states: A__ : List[Any] =all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : List[str] = PoolFormerConfig __magic_name__ : int = """poolformer""" __magic_name__ : Any = """pixel_values""" __magic_name__ : Any = True def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : str ): if isinstance(UpperCamelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any]=False ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A__ : Optional[Any] =value __A : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : Dict ): super().__init__(UpperCamelCase__ ) A__ : List[Any] =config A__ : Optional[Any] =PoolFormerEncoder(UpperCamelCase__ ) # Initialize weights and apply final processing self.post_init() def _UpperCAmelCase ( self : Tuple ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : int =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A__ : Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) A__ : List[Any] =self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : int =encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[Any] ): super().__init__() A__ : List[str] =nn.Linear(config.hidden_size , config.hidden_size ) def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] ): A__ : int =self.dense(UpperCamelCase__ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : str ): super().__init__(UpperCamelCase__ ) A__ : List[str] =config.num_labels A__ : Optional[int] =PoolFormerModel(UpperCamelCase__ ) # Final norm A__ : Dict =PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head A__ : Dict =( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.LongTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : Tuple =return_dict if return_dict is not None else self.config.use_return_dict A__ : List[str] =self.poolformer( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : str =outputs[0] A__ : List[Any] =self.classifier(self.norm(UpperCamelCase__ ).mean([-2, -1] ) ) A__ : Optional[Any] =None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: A__ : int ="regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): A__ : Tuple ="single_label_classification" else: A__ : Optional[int] ="multi_label_classification" if self.config.problem_type == "regression": A__ : Dict =MSELoss() if self.num_labels == 1: A__ : Optional[Any] =loss_fct(logits.squeeze() , labels.squeeze() ) else: A__ : List[str] =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": A__ : Tuple =CrossEntropyLoss() A__ : int =loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": A__ : List[Any] =BCEWithLogitsLoss() A__ : str =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: A__ : Optional[int] =(logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states )
656
0
def a_ ( __lowercase : list[list[int]] , __lowercase : int , __lowercase : int , __lowercase : set ) -> Any: _snake_case = len(__lowercase ), len(grid[0] ) if ( min(__lowercase , __lowercase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) _snake_case = 0 count += depth_first_search(__lowercase , row + 1 , __lowercase , __lowercase ) count += depth_first_search(__lowercase , row - 1 , __lowercase , __lowercase ) count += depth_first_search(__lowercase , __lowercase , col + 1 , __lowercase ) count += depth_first_search(__lowercase , __lowercase , col - 1 , __lowercase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
686
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : int = IFInpaintingSuperResolutionPipeline __magic_name__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __magic_name__ : str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""}) __magic_name__ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def _UpperCAmelCase ( self : Union[str, Any] ): return self._get_superresolution_dummy_components() def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[int]=0 ): if str(UpperCamelCase__ ).startswith("mps" ): A__ : Any =torch.manual_seed(UpperCamelCase__ ) else: A__ : Dict =torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A__ : Tuple =floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Optional[int] =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Any =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : List[str] ={ "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _UpperCAmelCase ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _UpperCAmelCase ( self : int ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _UpperCAmelCase ( self : Tuple ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def _UpperCAmelCase ( self : str ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _UpperCAmelCase ( self : Dict ): self._test_save_load_local() def _UpperCAmelCase ( self : Optional[int] ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
656
0
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow A : Optional[int] = False class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self , __a=32 ): set_seed(0 ) __lowerCAmelCase = UNetaDModel(sample_size=UpperCamelCase__ , in_channels=3 , out_channels=3 ) __lowerCAmelCase = torch.optim.SGD(model.parameters() , lr=0.0_0_0_1 ) return model, optimizer @slow def snake_case ( self ): __lowerCAmelCase = "cpu" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable __lowerCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_start=0.0_0_0_1 , beta_end=0.0_2 , beta_schedule="linear" , clip_sample=UpperCamelCase__ , ) __lowerCAmelCase = DDIMScheduler( num_train_timesteps=10_00 , beta_start=0.0_0_0_1 , beta_end=0.0_2 , beta_schedule="linear" , clip_sample=UpperCamelCase__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) __lowerCAmelCase = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(UpperCamelCase__ ) for _ in range(4 )] __lowerCAmelCase = [torch.randn((4, 3, 32, 32) ).to(UpperCamelCase__ ) for _ in range(4 )] __lowerCAmelCase = [torch.randint(0 , 10_00 , (4,) ).long().to(UpperCamelCase__ ) for _ in range(4 )] # train with a DDPM scheduler __lowerCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(UpperCamelCase__ ) for i in range(4 ): optimizer.zero_grad() __lowerCAmelCase = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) __lowerCAmelCase = model(UpperCamelCase__ , timesteps[i] ).sample __lowerCAmelCase = torch.nn.functional.mse_loss(UpperCamelCase__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM __lowerCAmelCase = self.get_model_optimizer(resolution=32 ) model.train().to(UpperCamelCase__ ) for i in range(4 ): optimizer.zero_grad() __lowerCAmelCase = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) __lowerCAmelCase = model(UpperCamelCase__ , timesteps[i] ).sample __lowerCAmelCase = torch.nn.functional.mse_loss(UpperCamelCase__ , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-5 ) ) self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-5 ) )
636
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __A : Any = { "configuration_efficientformer": [ "EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientFormerConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = ["EfficientFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientFormerForImageClassification", "EfficientFormerForImageClassificationWithTeacher", "EfficientFormerModel", "EfficientFormerPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[int] = [ "TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFEfficientFormerForImageClassification", "TFEfficientFormerForImageClassificationWithTeacher", "TFEfficientFormerModel", "TFEfficientFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys __A : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { "facebook/wav2vec2-base-960h": "https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json", # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class lowerCAmelCase_ ( _UpperCamelCase ): '''simple docstring''' _lowerCamelCase: Optional[Any] = """wav2vec2""" def __init__( self : Dict ,A_ : Union[str, Any]=32 ,A_ : int=768 ,A_ : Union[str, Any]=12 ,A_ : Optional[int]=12 ,A_ : List[Any]=3072 ,A_ : int="gelu" ,A_ : str=0.1 ,A_ : int=0.1 ,A_ : Optional[Any]=0.1 ,A_ : List[Any]=0.0 ,A_ : Optional[int]=0.0 ,A_ : Optional[int]=0.1 ,A_ : Any=0.1 ,A_ : Dict=0.02 ,A_ : List[Any]=1e-5 ,A_ : Dict="group" ,A_ : Any="gelu" ,A_ : Optional[int]=(512, 512, 512, 512, 512, 512, 512) ,A_ : Optional[int]=(5, 2, 2, 2, 2, 2, 2) ,A_ : Optional[int]=(10, 3, 3, 3, 3, 2, 2) ,A_ : Tuple=False ,A_ : Optional[Any]=128 ,A_ : Optional[int]=16 ,A_ : Optional[Any]=False ,A_ : List[Any]=True ,A_ : List[Any]=0.05 ,A_ : Optional[int]=10 ,A_ : Optional[int]=2 ,A_ : Any=0.0 ,A_ : Union[str, Any]=10 ,A_ : Optional[Any]=0 ,A_ : str=320 ,A_ : Optional[Any]=2 ,A_ : Optional[Any]=0.1 ,A_ : int=100 ,A_ : str=256 ,A_ : Optional[int]=256 ,A_ : str=0.1 ,A_ : Optional[int]="sum" ,A_ : str=False ,A_ : List[Any]=False ,A_ : Optional[Any]=256 ,A_ : Any=(512, 512, 512, 512, 1500) ,A_ : Tuple=(5, 3, 3, 1, 1) ,A_ : Optional[Any]=(1, 2, 3, 1, 1) ,A_ : Tuple=512 ,A_ : Dict=0 ,A_ : int=1 ,A_ : List[Any]=2 ,A_ : List[str]=False ,A_ : List[str]=3 ,A_ : List[str]=2 ,A_ : List[Any]=3 ,A_ : List[Any]=None ,A_ : int=None ,**A_ : List[Any] ,) -> Optional[int]: super().__init__(**UpperCamelCase__ ,pad_token_id=UpperCamelCase__ ,bos_token_id=UpperCamelCase__ ,eos_token_id=UpperCamelCase__ ) A = hidden_size A = feat_extract_norm A = feat_extract_activation A = list(UpperCamelCase__ ) A = list(UpperCamelCase__ ) A = list(UpperCamelCase__ ) A = conv_bias A = num_conv_pos_embeddings A = num_conv_pos_embedding_groups A = len(self.conv_dim ) A = num_hidden_layers A = intermediate_size A = hidden_act A = num_attention_heads A = hidden_dropout A = attention_dropout A = activation_dropout A = feat_proj_dropout A = final_dropout A = layerdrop A = layer_norm_eps A = initializer_range A = vocab_size A = do_stable_layer_norm A = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' F' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,' F' `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 A = apply_spec_augment A = mask_time_prob A = mask_time_length A = mask_time_min_masks A = mask_feature_prob A = mask_feature_length A = mask_feature_min_masks # parameters for pretraining with codevector quantized representations A = num_codevectors_per_group A = num_codevector_groups A = contrastive_logits_temperature A = feat_quantizer_dropout A = num_negatives A = codevector_dim A = proj_codevector_dim A = diversity_loss_weight # ctc loss A = ctc_loss_reduction A = ctc_zero_infinity # adapter A = add_adapter A = adapter_kernel_size A = adapter_stride A = num_adapter_layers A = output_hidden_size or hidden_size A = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. A = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. A = list(UpperCamelCase__ ) A = list(UpperCamelCase__ ) A = list(UpperCamelCase__ ) A = xvector_output_dim @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: return functools.reduce(operator.mul ,self.conv_stride ,1 )
91
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Tuple =[] for _ in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Union[str, Any]=10 ): """simple docstring""" A__ : Dict =[] for step in range(UpperCamelCase ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] =os.path.join(UpperCamelCase , "schedule.bin" ) torch.save(scheduler.state_dict() , UpperCamelCase ) A__ : Dict =torch.load(UpperCamelCase ) scheduler.load_state_dict(UpperCamelCase ) return lrs @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : Any =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Optional[Any] =torch.tensor([0.4, 0.2, -0.5] ) A__ : Any =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : List[str] =AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(100 ): A__ : Optional[int] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def _UpperCAmelCase ( self : Dict ): A__ : Optional[int] =torch.tensor([0.1, -0.2, -0.1] , requires_grad=UpperCamelCase__ ) A__ : Dict =torch.tensor([0.4, 0.2, -0.5] ) A__ : Optional[int] =nn.MSELoss() # No warmup, constant schedule, no gradient clipping A__ : int =Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=UpperCamelCase__ , weight_decay=0.0 , relative_step=UpperCamelCase__ , scale_parameter=UpperCamelCase__ , warmup_init=UpperCamelCase__ , ) for _ in range(1000 ): A__ : List[Any] =criterion(UpperCamelCase__ , UpperCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' __magic_name__ : Optional[int] = nn.Linear(50 , 50) if is_torch_available() else None __magic_name__ : Any = AdamW(m.parameters() , lr=10.0) if is_torch_available() else None __magic_name__ : Union[str, Any] = 10 def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for a, b in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertAlmostEqual(UpperCamelCase__ , UpperCamelCase__ , delta=UpperCamelCase__ , msg=UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[Any] ): A__ : Union[str, Any] ={"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) A__ : Union[str, Any] ={ get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1E-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): A__ , A__ : Any =data A__ : Union[str, Any] =scheduler_func(self.optimizer , **UpperCamelCase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) A__ : int =unwrap_schedule(UpperCamelCase__ , self.num_steps ) self.assertListAlmostEqual( UpperCamelCase__ , UpperCamelCase__ , tol=1E-2 , msg=F'''failed for {scheduler_func} in normal scheduler''' , ) A__ : List[str] =scheduler_func(self.optimizer , **UpperCamelCase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(UpperCamelCase__ ) # wrap to test picklability of the schedule A__ : Tuple =unwrap_and_save_reload_schedule(UpperCamelCase__ , self.num_steps ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ , msg=F'''failed for {scheduler_func} in save and reload''' ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : str ): A__ : int =fn def __call__( self : List[Any] , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : List[Any] ): return self.fn(*UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict ): A__ : str =list(map(self , scheduler.lr_lambdas ) )
656
0
"""simple docstring""" def _UpperCamelCase ( A ): return credit_card_number.startswith(("34", "35", "37", "4", "5", "6") ) def _UpperCamelCase ( A ): UpperCamelCase_ =credit_card_number UpperCamelCase_ =0 UpperCamelCase_ =len(A ) - 2 for i in range(A , -1 , -2 ): # double the value of every second digit UpperCamelCase_ =int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 UpperCamelCase_ =cc_number[:i] + str(A ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(A ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def _UpperCamelCase ( A ): UpperCamelCase_ =f"""{credit_card_number} is an invalid credit card number because""" if not credit_card_number.isdigit(): print(f"""{error_message} it has nonnumerical characters.""" ) return False if not 13 <= len(A ) <= 16: print(f"""{error_message} of its length.""" ) return False if not validate_initial_digits(A ): print(f"""{error_message} of its first two digits.""" ) return False if not luhn_validation(A ): print(f"""{error_message} it fails the Luhn check.""" ) return False print(f"""{credit_card_number} is a valid credit card number.""" ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("4111111111111111") validate_credit_card_number("32323")
391
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __A : List[Any] = logging.get_logger("transformers.models.speecht5") __A : Optional[Any] = { "speech_encoder_prenet.layer_norm": "speecht5.encoder.prenet.feature_projection.layer_norm", "speech_encoder_prenet.post_extract_proj": "speecht5.encoder.prenet.feature_projection.projection", "speech_encoder_prenet.pos_conv.0": "speecht5.encoder.prenet.pos_conv_embed.conv", "speech_encoder_prenet.mask_emb": "speecht5.encoder.prenet.masked_spec_embed", } __A : Optional[int] = { "text_encoder_prenet.encoder_prenet.0": "speecht5.encoder.prenet.embed_tokens", "text_encoder_prenet.encoder_prenet.1.alpha": "speecht5.encoder.prenet.encode_positions.alpha", } __A : List[str] = { "speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0": "speecht5.decoder.prenet.layers.0", "speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0": "speecht5.decoder.prenet.layers.1", "speech_decoder_prenet.decoder_prenet.0.1": "speecht5.decoder.prenet.final_layer", "speech_decoder_prenet.decoder_prenet.1.alpha": "speecht5.decoder.prenet.encode_positions.alpha", "speech_decoder_prenet.spkembs_layer.0": "speecht5.decoder.prenet.speaker_embeds_layer", } __A : List[Any] = { "speech_decoder_postnet.feat_out": "speech_decoder_postnet.feat_out", "speech_decoder_postnet.prob_out": "speech_decoder_postnet.prob_out", "speech_decoder_postnet.postnet.postnet.0.0": "speech_decoder_postnet.layers.0.conv", "speech_decoder_postnet.postnet.postnet.0.1": "speech_decoder_postnet.layers.0.batch_norm", "speech_decoder_postnet.postnet.postnet.1.0": "speech_decoder_postnet.layers.1.conv", "speech_decoder_postnet.postnet.postnet.1.1": "speech_decoder_postnet.layers.1.batch_norm", "speech_decoder_postnet.postnet.postnet.2.0": "speech_decoder_postnet.layers.2.conv", "speech_decoder_postnet.postnet.postnet.2.1": "speech_decoder_postnet.layers.2.batch_norm", "speech_decoder_postnet.postnet.postnet.3.0": "speech_decoder_postnet.layers.3.conv", "speech_decoder_postnet.postnet.postnet.3.1": "speech_decoder_postnet.layers.3.batch_norm", "speech_decoder_postnet.postnet.postnet.4.0": "speech_decoder_postnet.layers.4.conv", "speech_decoder_postnet.postnet.postnet.4.1": "speech_decoder_postnet.layers.4.batch_norm", } __A : Union[str, Any] = { "text_decoder_prenet.embed_tokens": "speecht5.decoder.prenet.embed_tokens", } __A : Any = { "text_decoder_postnet.output_projection": "text_decoder_postnet.lm_head", } __A : Union[str, Any] = { "encoder.layers.*.self_attn.k_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj", "encoder.layers.*.self_attn.v_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj", "encoder.layers.*.self_attn.q_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj", "encoder.layers.*.self_attn.out_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj", "encoder.layers.*.self_attn_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.layer_norm", "encoder.layers.*.fc1": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense", "encoder.layers.*.fc2": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense", "encoder.layers.*.final_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm", "encoder.layer_norm": "speecht5.encoder.wrapped_encoder.layer_norm", "encoder.pos_emb.pe_k": "speecht5.encoder.wrapped_encoder.embed_positions.pe_k", } __A : Optional[int] = { "decoder.layers.*.self_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj", "decoder.layers.*.self_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj", "decoder.layers.*.self_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj", "decoder.layers.*.self_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj", "decoder.layers.*.self_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm", "decoder.layers.*.encoder_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj", "decoder.layers.*.encoder_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj", "decoder.layers.*.encoder_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj", "decoder.layers.*.encoder_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj", "decoder.layers.*.encoder_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm", "decoder.layers.*.fc1": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense", "decoder.layers.*.fc2": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense", "decoder.layers.*.final_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm", } __A : Union[str, Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __A : Optional[Any] = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : Optional[int] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __A : int = [] __A : int = [ "encoder.version", "encoder.layers.*.norm_k.weight", "encoder.layers.*.norm_k.bias", "decoder.version", "decoder.layers.*.norm_k.weight", "decoder.layers.*.norm_k.bias", "decoder.pos_emb.pe_k", "speech_encoder_prenet.embed_positions._float_tensor", "text_decoder_prenet.embed_positions._float_tensor", ] __A : Optional[Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "speech_decoder_prenet.*", "speech_decoder_postnet.*", ] __A : Tuple = IGNORE_KEYS + [ "encoder.proj", "speech_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] __A : Union[str, Any] = IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] def lowercase ( UpperCamelCase : str , UpperCamelCase : Union[str, Any] , UpperCamelCase : int , UpperCamelCase : List[Any] , UpperCamelCase : int ): """simple docstring""" for attribute in key.split("." ): A__ : Dict =getattr(UpperCamelCase , UpperCamelCase ) if weight_type is not None: A__ : Union[str, Any] =getattr(UpperCamelCase , UpperCamelCase ).shape else: A__ : Tuple =hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": A__ : Any =value elif weight_type == "weight_g": A__ : Any =value elif weight_type == "weight_v": A__ : Any =value elif weight_type == "bias": A__ : Tuple =value elif weight_type == "running_mean": A__ : Dict =value elif weight_type == "running_var": A__ : List[str] =value elif weight_type == "num_batches_tracked": A__ : Dict =value else: A__ : Optional[int] =value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''' ) def lowercase ( UpperCamelCase : Tuple , UpperCamelCase : Tuple ): """simple docstring""" for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: A__ , A__ : List[str] =key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : Dict ): """simple docstring""" A__ : Tuple =[] if task == "s2t": A__ : Dict =hf_model.speechta.encoder.prenet.feature_encoder A__ : int =MAPPING_S2T A__ : List[Any] =IGNORE_KEYS_S2T elif task == "t2s": A__ : Union[str, Any] =None A__ : List[Any] =MAPPING_T2S A__ : Tuple =IGNORE_KEYS_T2S elif task == "s2s": A__ : Optional[Any] =hf_model.speechta.encoder.prenet.feature_encoder A__ : Tuple =MAPPING_S2S A__ : Any =IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''' ) for name, value in fairseq_dict.items(): if should_ignore(UpperCamelCase , UpperCamelCase ): logger.info(F'''{name} was ignored''' ) continue A__ : Optional[Any] =False if "conv_layers" in name: load_conv_layer( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , hf_model.config.feat_extract_norm == "group" , ) A__ : List[Any] =True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: A__ , A__ : Dict =key.split(".*." ) if prefix in name and suffix in name: A__ : int =suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: A__ : List[Any] =True if "*" in mapped_key: A__ : Optional[int] =name.split(UpperCamelCase )[0].split("." )[-2] A__ : int =mapped_key.replace("*" , UpperCamelCase ) if "weight_g" in name: A__ : str ="weight_g" elif "weight_v" in name: A__ : Optional[Any] ="weight_v" elif "bias" in name: A__ : Any ="bias" elif "weight" in name: A__ : Optional[int] ="weight" elif "running_mean" in name: A__ : Tuple ="running_mean" elif "running_var" in name: A__ : Optional[int] ="running_var" elif "num_batches_tracked" in name: A__ : str ="num_batches_tracked" else: A__ : List[Any] =None set_recursively(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) continue if not is_used: unused_weights.append(UpperCamelCase ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Dict , UpperCamelCase : Any , UpperCamelCase : str , UpperCamelCase : Dict ): """simple docstring""" A__ : Any =full_name.split("conv_layers." )[-1] A__ : Dict =name.split("." ) A__ : int =int(items[0] ) A__ : str =int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) A__ : Optional[Any] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) A__ : Optional[int] =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''' ) A__ : Any =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(UpperCamelCase ) @torch.no_grad() def lowercase ( UpperCamelCase : Any , UpperCamelCase : Union[str, Any] , UpperCamelCase : List[str] , UpperCamelCase : str=None , UpperCamelCase : Any=None , UpperCamelCase : Tuple=None , ): """simple docstring""" if config_path is not None: A__ : Any =SpeechTaConfig.from_pretrained(UpperCamelCase ) else: A__ : Any =SpeechTaConfig() if task == "s2t": A__ : Union[str, Any] =config.max_text_positions A__ : Dict =SpeechTaForSpeechToText(UpperCamelCase ) elif task == "t2s": A__ : str =1876 A__ : Optional[int] =600 A__ : Tuple =config.max_speech_positions A__ : Optional[Any] =SpeechTaForTextToSpeech(UpperCamelCase ) elif task == "s2s": A__ : str =1876 A__ : Tuple =config.max_speech_positions A__ : Any =SpeechTaForSpeechToSpeech(UpperCamelCase ) else: raise ValueError(F'''Unknown task name: {task}''' ) if vocab_path: A__ : str =SpeechTaTokenizer(UpperCamelCase , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it A__ : Optional[Any] =AddedToken("<mask>" , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) A__ : int =mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) A__ : Dict =SpeechTaFeatureExtractor() A__ : Tuple =SpeechTaProcessor(tokenizer=UpperCamelCase , feature_extractor=UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) A__ : Union[str, Any] =torch.load(UpperCamelCase ) recursively_load_weights(fairseq_checkpoint["model"] , UpperCamelCase , UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if repo_id: print("Pushing to the hub..." ) processor.push_to_hub(UpperCamelCase ) model.push_to_hub(UpperCamelCase ) if __name__ == "__main__": __A : Dict = argparse.ArgumentParser() parser.add_argument( "--task", default="s2t", type=str, help="Type of the SpeechT5 model you'd like to convert. Should be one of 's2t', 't2s', 's2s'.", ) parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--vocab_path", default=None, type=str, help="Path to SentencePiece model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) __A : str = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
656
0
'''simple docstring''' import comet # From: unbabel-comet import torch import datasets __SCREAMING_SNAKE_CASE : List[Any] = datasets.logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Any = "\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unbabel's Participation in the WMT20 Metrics Shared Task},\n booktitle = {Proceedings of the Fifth Conference on Machine Translation},\n month = {November},\n year = {2020},\n address = {Online},\n publisher = {Association for Computational Linguistics},\n pages = {909--918},\n}\n@inproceedings{rei-etal-2020-comet,\n title = \"{COMET}: A Neural Framework for {MT} Evaluation\",\n author = \"Rei, Ricardo and\n Stewart, Craig and\n Farinha, Ana C and\n Lavie, Alon\",\n booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\",\n month = nov,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\",\n pages = \"2685--2702\",\n}\n" __SCREAMING_SNAKE_CASE : Optional[int] = "\\nCrosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM).\nWith the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition.\n\nSee the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information.\n" __SCREAMING_SNAKE_CASE : Any = "\nCOMET score.\n\nArgs:\n\n`sources` (list of str): Source sentences\n`predictions` (list of str): candidate translations\n`references` (list of str): reference translations\n`cuda` (bool): If set to True, runs COMET using GPU\n`show_progress` (bool): Shows progress\n`model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None.\n\nReturns:\n `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`.\n `scores`: List of scores.\n\nExamples:\n\n >>> comet_metric = datasets.load_metric('comet')\n >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use\n >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"]\n >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"]\n >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"]\n >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source)\n >>> print([round(v, 2) for v in results[\"scores\"]])\n [0.19, 0.92]\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): """simple docstring""" def _UpperCAmelCase ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence" ), "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def _UpperCAmelCase ( self : str , lowerCAmelCase : List[Any] ): if self.config_name == "default": A_ = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da" ) ) else: A_ = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _UpperCAmelCase ( self : int , lowerCAmelCase : Optional[int] , lowerCAmelCase : Optional[int] , lowerCAmelCase : Any , lowerCAmelCase : Dict=None , lowerCAmelCase : str=False ): if gpus is None: A_ = 1 if torch.cuda.is_available() else 0 A_ = {"src": sources, "mt": predictions, "ref": references} A_ = [dict(zip(UpperCamelCase__ , UpperCamelCase__ ) ) for t in zip(*data.values() )] A_ = self.scorer.predict(UpperCamelCase__ , gpus=UpperCamelCase__ , progress_bar=UpperCamelCase__ ) return {"mean_score": mean_score, "scores": scores}
452
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase): '''simple docstring''' __magic_name__ : List[Any] = [R"""h\.\d+\.attn\.bias""", R"""h\.\d+\.attn\.masked_bias"""] @register_to_config def __init__( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : int = 50257 , UpperCamelCase__ : int = 1024 , UpperCamelCase__ : int = 768 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : int = 12 , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : str = "gelu_new" , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 0.1 , UpperCamelCase__ : float = 1E-5 , UpperCamelCase__ : float = 0.02 , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = True , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , ): super().__init__() A__ : Dict =prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) A__ : Optional[int] =prefix_inner_dim A__ : Optional[int] =prefix_hidden_dim A__ : Optional[int] =( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : Optional[int] =( nn.Linear(self.prefix_hidden_dim , UpperCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) A__ : str =GPTaConfig( vocab_size=UpperCamelCase__ , n_positions=UpperCamelCase__ , n_embd=UpperCamelCase__ , n_layer=UpperCamelCase__ , n_head=UpperCamelCase__ , n_inner=UpperCamelCase__ , activation_function=UpperCamelCase__ , resid_pdrop=UpperCamelCase__ , embd_pdrop=UpperCamelCase__ , attn_pdrop=UpperCamelCase__ , layer_norm_epsilon=UpperCamelCase__ , initializer_range=UpperCamelCase__ , scale_attn_weights=UpperCamelCase__ , use_cache=UpperCamelCase__ , scale_attn_by_inverse_layer_idx=UpperCamelCase__ , reorder_and_upcast_attn=UpperCamelCase__ , ) A__ : Any =GPTaLMHeadModel(UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : torch.Tensor , UpperCamelCase__ : Optional[torch.Tensor] = None , UpperCamelCase__ : Optional[torch.Tensor] = None , ): A__ : int =self.transformer.transformer.wte(UpperCamelCase__ ) A__ : Tuple =self.encode_prefix(UpperCamelCase__ ) A__ : Union[str, Any] =self.decode_prefix(UpperCamelCase__ ) A__ : Tuple =torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A__ : Any =self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A__ : List[Any] =torch.cat((dummy_token, input_ids) , dim=1 ) A__ : Any =self.transformer(inputs_embeds=UpperCamelCase__ , labels=UpperCamelCase__ , attention_mask=UpperCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : torch.device ): return torch.zeros(UpperCamelCase__ , self.prefix_length , dtype=torch.intaa , device=UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Tuple ): return self.encode_prefix(UpperCamelCase__ ) @torch.no_grad() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : str ): A__ : Optional[int] =torch.split(UpperCamelCase__ , 1 , dim=0 ) A__ : List[str] =[] A__ : Dict =[] for feature in features: A__ : Any =self.decode_prefix(feature.to(UpperCamelCase__ ) ) # back to the clip feature # Only support beam search for now A__ , A__ : Optional[Any] =self.generate_beam( input_embeds=UpperCamelCase__ , device=UpperCamelCase__ , eos_token_id=UpperCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A__ : Optional[Any] =torch.stack(UpperCamelCase__ ) A__ : Optional[int] =torch.stack(UpperCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _UpperCAmelCase ( self : List[Any] , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : List[str]=None , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : int = 5 , UpperCamelCase__ : int = 67 , UpperCamelCase__ : float = 1.0 , UpperCamelCase__ : Optional[int] = None , ): A__ : str =eos_token_id A__ : Optional[Any] =None A__ : int =None A__ : Union[str, Any] =torch.ones(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.int ) A__ : Any =torch.zeros(UpperCamelCase__ , device=UpperCamelCase__ , dtype=torch.bool ) if input_embeds is not None: A__ : Union[str, Any] =input_embeds else: A__ : Optional[Any] =self.transformer.transformer.wte(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): A__ : Optional[int] =self.transformer(inputs_embeds=UpperCamelCase__ ) A__ : Tuple =outputs.logits A__ : Union[str, Any] =logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A__ : Optional[Any] =logits.softmax(-1 ).log() if scores is None: A__ , A__ : Union[str, Any] =logits.topk(UpperCamelCase__ , -1 ) A__ : Union[str, Any] =generated.expand(UpperCamelCase__ , *generated.shape[1:] ) A__ , A__ : Optional[int] =next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A__ : str =next_tokens else: A__ : Optional[Any] =tokens.expand(UpperCamelCase__ , *tokens.shape[1:] ) A__ : str =torch.cat((tokens, next_tokens) , dim=1 ) else: A__ : Union[str, Any] =-float(np.inf ) A__ : Dict =0 A__ : Optional[Any] =scores[:, None] + logits seq_lengths[~is_stopped] += 1 A__ : Optional[Any] =scores_sum / seq_lengths[:, None] A__ , A__ : List[Any] =scores_sum_average.view(-1 ).topk(UpperCamelCase__ , -1 ) A__ : Tuple =next_tokens // scores_sum.shape[1] A__ : List[Any] =seq_lengths[next_tokens_source] A__ : int =next_tokens % scores_sum.shape[1] A__ : str =next_tokens.unsqueeze(1 ) A__ : List[Any] =tokens[next_tokens_source] A__ : int =torch.cat((tokens, next_tokens) , dim=1 ) A__ : List[str] =generated[next_tokens_source] A__ : Optional[Any] =scores_sum_average * seq_lengths A__ : Optional[int] =is_stopped[next_tokens_source] A__ : List[str] =self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A__ : str =torch.cat((generated, next_token_embed) , dim=1 ) A__ : str =is_stopped + next_tokens.eq(UpperCamelCase__ ).squeeze() if is_stopped.all(): break A__ : Optional[int] =scores / seq_lengths A__ : List[Any] =scores.argsort(descending=UpperCamelCase__ ) # tokens tensors are already padded to max_seq_length A__ : int =[tokens[i] for i in order] A__ : Any =torch.stack(UpperCamelCase__ , dim=0 ) A__ : int =torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
656
0
"""simple docstring""" import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __snake_case (_UpperCamelCase , unittest.TestCase ): __a = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def __a ( self: Dict , A_: Dict=0 ): __lowerCamelCase = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(UpperCamelCase__ ) ) __lowerCamelCase = np.random.RandomState(UpperCamelCase__ ) __lowerCamelCase = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "strength": 0.75, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def __a ( self: Dict ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __a ( self: Dict ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __lowerCamelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __a ( self: Optional[int] ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __lowerCamelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) # warmup pass to apply optimizations __lowerCamelCase = pipe(**self.get_dummy_inputs() ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __a ( self: str ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __lowerCamelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __a ( self: int ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __lowerCamelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __a ( self: Any ): __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __lowerCamelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**UpperCamelCase__ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __lowerCamelCase = np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __snake_case (unittest.TestCase ): @property def __a ( self: Tuple ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __a ( self: Any ): __lowerCamelCase = ort.SessionOptions() __lowerCamelCase = False return options def __a ( self: Any ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __lowerCamelCase = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = "A fantasy landscape, trending on artstation" __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type="""np""" , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) __lowerCamelCase = np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __a ( self: List[str] ): __lowerCamelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __lowerCamelCase = init_image.resize((7_68, 5_12) ) __lowerCamelCase = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowerCamelCase = "A fantasy landscape, trending on artstation" __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=UpperCamelCase__ , image=UpperCamelCase__ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCamelCase__ , output_type="""np""" , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) __lowerCamelCase = np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
281
"""simple docstring""" import os def lowercase ( ): """simple docstring""" A__ : List[Any] =os.path.dirname(os.path.realpath(UpperCamelCase ) ) A__ : str =os.path.join(UpperCamelCase , "triangle.txt" ) with open(UpperCamelCase ) as f: A__ : Optional[int] =f.readlines() A__ : str =[] for line in triangle: A__ : Union[str, Any] =[] for number in line.strip().split(" " ): numbers_from_line.append(int(UpperCamelCase ) ) a.append(UpperCamelCase ) for i in range(1 , len(UpperCamelCase ) ): for j in range(len(a[i] ) ): A__ : Union[str, Any] =a[i - 1][j] if j != len(a[i - 1] ) else 0 A__ : Union[str, Any] =a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(UpperCamelCase , UpperCamelCase ) return max(a[-1] ) if __name__ == "__main__": print(solution())
656
0
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase : Any =get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _UpperCamelCase : Tuple =250004 _UpperCamelCase : Optional[int] =250020 @require_sentencepiece @require_tokenizers class UpperCAmelCase__ ( _UpperCamelCase , unittest.TestCase ): __snake_case : List[Any] = MBartTokenizer __snake_case : Dict = MBartTokenizerFast __snake_case : Tuple = True __snake_case : int = True def A__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _A : str = MBartTokenizer(UpperCamelCase__ ,keep_accents=UpperCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self ): _A : str = MBartTokenizer(UpperCamelCase__ ,keep_accents=UpperCamelCase__ ) _A : List[Any] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCamelCase__ ,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) _A : Optional[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCamelCase__ ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] ,) _A : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] ,) _A : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] ,) def A__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _A : Any = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _A : Any = self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ ,**UpperCamelCase__ ) _A : List[str] = self.tokenizer_class.from_pretrained(UpperCamelCase__ ,**UpperCamelCase__ ) _A : Any = tempfile.mkdtemp() _A : str = tokenizer_r.save_pretrained(UpperCamelCase__ ) _A : int = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _A : Tuple = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(UpperCamelCase__ ,UpperCamelCase__ ) # Checks everything loads correctly in the same way _A : Dict = tokenizer_r.from_pretrained(UpperCamelCase__ ) _A : List[Any] = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ ,UpperCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase__ ) # Save tokenizer rust, legacy_format=True _A : Optional[Any] = tempfile.mkdtemp() _A : str = tokenizer_r.save_pretrained(UpperCamelCase__ ,legacy_format=UpperCamelCase__ ) _A : List[str] = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase__ ,UpperCamelCase__ ) # Checks everything loads correctly in the same way _A : int = tokenizer_r.from_pretrained(UpperCamelCase__ ) _A : int = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ ,UpperCamelCase__ ) ) shutil.rmtree(UpperCamelCase__ ) # Save tokenizer rust, legacy_format=False _A : List[str] = tempfile.mkdtemp() _A : str = tokenizer_r.save_pretrained(UpperCamelCase__ ,legacy_format=UpperCamelCase__ ) _A : Any = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _A : List[str] = tokenizer_r.from_pretrained(UpperCamelCase__ ) _A : Any = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ ,UpperCamelCase__ ) ) shutil.rmtree(UpperCamelCase__ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase__ ( unittest.TestCase ): __snake_case : Dict = """facebook/mbart-large-en-ro""" __snake_case : List[Any] = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] __snake_case : str = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] __snake_case : Dict = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def A__ ( cls ): _A : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name ,src_lang='''en_XX''' ,tgt_lang='''ro_RO''' ) _A : List[Any] = 1 return cls def A__ ( self ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] ,250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] ,250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] ,250020 ) def A__ ( self ): _A : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens ,UpperCamelCase__ ) def A__ ( self ): self.assertIn(UpperCamelCase__ ,self.tokenizer.all_special_ids ) _A : Optional[Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] _A : Tuple = self.tokenizer.decode(UpperCamelCase__ ,skip_special_tokens=UpperCamelCase__ ) _A : List[Any] = self.tokenizer.decode(generated_ids[1:] ,skip_special_tokens=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ ,UpperCamelCase__ ) self.assertNotIn(self.tokenizer.eos_token ,UpperCamelCase__ ) def A__ ( self ): _A : Dict = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] ,UpperCamelCase__ ) _A : Any = 10 _A : Optional[Any] = self.tokenizer(UpperCamelCase__ ,max_length=UpperCamelCase__ ,truncation=UpperCamelCase__ ).input_ids[0] self.assertEqual(ids[-2] ,2 ) self.assertEqual(ids[-1] ,UpperCamelCase__ ) self.assertEqual(len(UpperCamelCase__ ) ,UpperCamelCase__ ) def A__ ( self ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) ,[250026, 250001] ) def A__ ( self ): _A : List[Any] = tempfile.mkdtemp() _A : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase__ ) _A : int = MBartTokenizer.from_pretrained(UpperCamelCase__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids ,UpperCamelCase__ ) @require_torch def A__ ( self ): _A : Optional[Any] = self.tokenizer(self.src_text ,text_target=self.tgt_text ,padding=UpperCamelCase__ ,return_tensors='''pt''' ) _A : str = shift_tokens_right(batch['''labels'''] ,self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def A__ ( self ): _A : Optional[int] = self.tokenizer( self.src_text ,text_target=self.tgt_text ,padding=UpperCamelCase__ ,truncation=UpperCamelCase__ ,max_length=len(self.expected_src_tokens ) ,return_tensors='''pt''' ,) _A : Union[str, Any] = shift_tokens_right(batch['''labels'''] ,self.tokenizer.pad_token_id ) self.assertIsInstance(UpperCamelCase__ ,UpperCamelCase__ ) self.assertEqual((2, 14) ,batch.input_ids.shape ) self.assertEqual((2, 14) ,batch.attention_mask.shape ) _A : Tuple = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens ,UpperCamelCase__ ) self.assertEqual(2 ,batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens ,[] ) self.assertEqual(self.tokenizer.suffix_tokens ,[self.tokenizer.eos_token_id, EN_CODE] ) def A__ ( self ): _A : Optional[Any] = self.tokenizer(self.src_text ,padding=UpperCamelCase__ ,truncation=UpperCamelCase__ ,max_length=3 ,return_tensors='''pt''' ) _A : Union[str, Any] = self.tokenizer( text_target=self.tgt_text ,padding=UpperCamelCase__ ,truncation=UpperCamelCase__ ,max_length=10 ,return_tensors='''pt''' ) _A : List[str] = targets["input_ids"] _A : int = shift_tokens_right(UpperCamelCase__ ,self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] ,3 ) self.assertEqual(batch.decoder_input_ids.shape[1] ,10 ) @require_torch def A__ ( self ): _A : Dict = self.tokenizer._build_translation_inputs( '''A test''' ,return_tensors='''pt''' ,src_lang='''en_XX''' ,tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(UpperCamelCase__ ) ,{ # A, test, EOS, en_XX '''input_ids''': [[62, 3034, 2, 250004]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 250001, } ,)
206
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A : int = logging.get_logger(__name__) def lowercase ( UpperCamelCase : Any ): """simple docstring""" A__ : str =OrderedDict() for key, value in state_dict.items(): if key.startswith("module.encoder" ): A__ : Dict =key.replace("module.encoder" , "glpn.encoder" ) if key.startswith("module.decoder" ): A__ : Optional[int] =key.replace("module.decoder" , "decoder.stages" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 A__ : Tuple =key[key.find("patch_embed" ) + len("patch_embed" )] A__ : Optional[Any] =key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(UpperCamelCase )-1}''' ) if "norm" in key: A__ : Dict =key.replace("norm" , "layer_norm" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 A__ : Any =key[key.find("glpn.encoder.layer_norm" ) + len("glpn.encoder.layer_norm" )] A__ : Tuple =key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(UpperCamelCase )-1}''' ) if "layer_norm1" in key: A__ : List[Any] =key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: A__ : Optional[int] =key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 A__ : int =key[key.find("block" ) + len("block" )] A__ : Optional[Any] =key.replace(F'''block{idx}''' , F'''block.{int(UpperCamelCase )-1}''' ) if "attn.q" in key: A__ : Optional[Any] =key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: A__ : Union[str, Any] =key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: A__ : str =key.replace("attn" , "attention.self" ) if "fc1" in key: A__ : Dict =key.replace("fc1" , "dense1" ) if "fc2" in key: A__ : str =key.replace("fc2" , "dense2" ) if "linear_pred" in key: A__ : List[Any] =key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: A__ : List[str] =key.replace("linear_fuse.conv" , "linear_fuse" ) A__ : Any =key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 A__ : str =key[key.find("linear_c" ) + len("linear_c" )] A__ : Dict =key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(UpperCamelCase )-1}''' ) if "bot_conv" in key: A__ : Union[str, Any] =key.replace("bot_conv" , "0.convolution" ) if "skip_conv1" in key: A__ : List[Any] =key.replace("skip_conv1" , "1.convolution" ) if "skip_conv2" in key: A__ : int =key.replace("skip_conv2" , "2.convolution" ) if "fusion1" in key: A__ : Optional[Any] =key.replace("fusion1" , "1.fusion" ) if "fusion2" in key: A__ : Optional[Any] =key.replace("fusion2" , "2.fusion" ) if "fusion3" in key: A__ : int =key.replace("fusion3" , "3.fusion" ) if "fusion" in key and "conv" in key: A__ : List[str] =key.replace("conv" , "convolutional_layer" ) if key.startswith("module.last_layer_depth" ): A__ : Tuple =key.replace("module.last_layer_depth" , "head.head" ) A__ : int =value return new_state_dict def lowercase ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Dict ): """simple docstring""" # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) A__ : int =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) A__ : str =state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict A__ : List[str] =kv_weight[ : config.hidden_sizes[i], : ] A__ : Dict =kv_bias[: config.hidden_sizes[i]] A__ : Any =kv_weight[ config.hidden_sizes[i] :, : ] A__ : Any =kv_bias[config.hidden_sizes[i] :] def lowercase ( ): """simple docstring""" A__ : Optional[Any] ="http://images.cocodataset.org/val2017/000000039769.jpg" A__ : List[Any] =Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return image @torch.no_grad() def lowercase ( UpperCamelCase : str , UpperCamelCase : Tuple , UpperCamelCase : List[str]=False , UpperCamelCase : str=None ): """simple docstring""" A__ : List[str] =GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) A__ : str =GLPNImageProcessor() # prepare image A__ : Any =prepare_img() A__ : Optional[int] =image_processor(images=UpperCamelCase , return_tensors="pt" ).pixel_values logger.info("Converting model..." ) # load original state dict A__ : int =torch.load(UpperCamelCase , map_location=torch.device("cpu" ) ) # rename keys A__ : Union[str, Any] =rename_keys(UpperCamelCase ) # key and value matrices need special treatment read_in_k_v(UpperCamelCase , UpperCamelCase ) # create HuggingFace model and load state dict A__ : Optional[int] =GLPNForDepthEstimation(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() # forward pass A__ : int =model(UpperCamelCase ) A__ : Optional[Any] =outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: A__ : List[Any] =torch.tensor( [[4.41_47, 4.08_73, 4.06_73], [3.78_90, 3.28_81, 3.15_25], [3.76_74, 3.54_23, 3.49_13]] ) elif "kitti" in model_name: A__ : Tuple =torch.tensor( [[3.42_91, 2.78_65, 2.51_51], [3.28_41, 2.70_21, 2.35_02], [3.11_47, 2.46_25, 2.24_81]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) A__ : str =torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , UpperCamelCase , atol=1E-4 ) print("Looks ok!" ) # finally, push to hub if required if push_to_hub: logger.info("Pushing model and image processor to the hub..." ) model.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=UpperCamelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCamelCase , UpperCamelCase ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=UpperCamelCase , ) if __name__ == "__main__": __A : List[str] = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) parser.add_argument( "--model_name", default="glpn-kitti", type=str, help="Name of the model in case you're pushing to the hub.", ) __A : Any = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
656
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __magic_name__ ( _UpperCamelCase ): def __init__( self : Optional[int] , snake_case_ : Union[str, Any] , snake_case_ : List[str] ): super().__init__() # make sure scheduler can always be converted to DDIM __snake_case = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=UpperCamelCase__ , scheduler=UpperCamelCase__ ) @torch.no_grad() def __call__( self : Dict , snake_case_ : int = 1 , snake_case_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , snake_case_ : float = 0.0 , snake_case_ : int = 50 , snake_case_ : Optional[bool] = None , snake_case_ : Optional[str] = "pil" , snake_case_ : bool = True , ): # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size , UpperCamelCase__ ): __snake_case = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: __snake_case = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(UpperCamelCase__ )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) __snake_case = randn_tensor(UpperCamelCase__ , generator=UpperCamelCase__ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(UpperCamelCase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __snake_case = self.unet(UpperCamelCase__ , UpperCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __snake_case = self.scheduler.step( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , eta=UpperCamelCase__ , use_clipped_model_output=UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample __snake_case = (image / 2 + 0.5).clamp(0 , 1 ) __snake_case = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case = self.numpy_to_pil(UpperCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCamelCase__ )
163
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging __A : Any = logging.get_logger(__name__) __A : Optional[Any] = { "EleutherAI/gpt-neo-1.3B": "https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Union[str, Any] = """gpt_neo""" __magic_name__ : Union[str, Any] = ["""past_key_values"""] __magic_name__ : Dict = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self : Dict , UpperCamelCase__ : List[Any]=50257 , UpperCamelCase__ : Optional[Any]=2048 , UpperCamelCase__ : Tuple=2048 , UpperCamelCase__ : int=24 , UpperCamelCase__ : Dict=[[["global", "local"], 12]] , UpperCamelCase__ : Optional[Any]=16 , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : str=256 , UpperCamelCase__ : List[str]="gelu_new" , UpperCamelCase__ : List[str]=0.0 , UpperCamelCase__ : Tuple=0.0 , UpperCamelCase__ : str=0.0 , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : List[str]=1E-5 , UpperCamelCase__ : Any=0.02 , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : Optional[Any]=50256 , UpperCamelCase__ : List[str]=50256 , **UpperCamelCase__ : str , ): A__ : Optional[Any] =vocab_size A__ : Dict =max_position_embeddings A__ : List[str] =hidden_size A__ : List[Any] =num_layers A__ : Tuple =num_heads A__ : List[str] =intermediate_size A__ : Tuple =window_size A__ : Dict =activation_function A__ : str =resid_dropout A__ : Union[str, Any] =embed_dropout A__ : List[str] =attention_dropout A__ : Tuple =classifier_dropout A__ : int =layer_norm_epsilon A__ : int =initializer_range A__ : str =use_cache A__ : Tuple =bos_token_id A__ : int =eos_token_id A__ : int =attention_types A__ : Any =self.expand_attention_types_params(UpperCamelCase__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.attention_layers)` == `config.num_layers` " F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' "`config.attention_layers` is prepared using `config.attention_types`. " "Please verify the value of `config.attention_types` argument." ) super().__init__(bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) @staticmethod def _UpperCAmelCase ( UpperCamelCase__ : List[str] ): A__ : Optional[Any] =[] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowercase ( UpperCamelCase : List[str] , UpperCamelCase : Optional[Any] , UpperCamelCase : List[Any] , UpperCamelCase : Optional[int] ): """simple docstring""" import torch A__ : List[str] =input.size() A__ : Dict =len(UpperCamelCase ) A__ : Optional[int] =shape[dimension] A__ : str =torch.arange(0 , UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =torch.div(sizedim - size , UpperCamelCase , rounding_mode="floor" ) + 1 A__ : str =torch.arange(UpperCamelCase ) + low_indices[:min_length][:, None] A__ : Tuple =[slice(UpperCamelCase )] * rank A__ : int =indices A__ : Optional[int] =input[s] A__ : Union[str, Any] =list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(UpperCamelCase ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any ): """simple docstring""" import torch A__ : List[str] =torch.arange(1 , UpperCamelCase ) A__ : List[Any] =torch.remainder(UpperCamelCase , UpperCamelCase ) A__ : Optional[int] =remainders == 0 A__ : str =candidates[divisor_indices] A__ : int =torch.max(UpperCamelCase ) return largest_divisor, torch.div(UpperCamelCase , UpperCamelCase , rounding_mode="floor" ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' @property def _UpperCAmelCase ( self : List[Any] ): A__ : Optional[int] =OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(UpperCamelCase__ , direction="inputs" ) A__ : Optional[int] ={0: "batch", 1: "past_sequence + sequence"} else: A__ : Tuple ={0: "batch", 1: "sequence"} return common_inputs @property def _UpperCAmelCase ( self : List[str] ): return self._config.num_heads def _UpperCAmelCase ( self : int , UpperCamelCase__ : PreTrainedTokenizer , UpperCamelCase__ : int = -1 , UpperCamelCase__ : int = -1 , UpperCamelCase__ : bool = False , UpperCamelCase__ : Optional[TensorType] = None , ): A__ : Union[str, Any] =super(UpperCamelCase__ , self ).generate_dummy_inputs( UpperCamelCase__ , batch_size=UpperCamelCase__ , seq_length=UpperCamelCase__ , is_pair=UpperCamelCase__ , framework=UpperCamelCase__ ) # We need to order the input in the way they appears in the forward() A__ : List[Any] =OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A__ , A__ : Union[str, Any] =common_inputs["input_ids"].shape # Not using the same length for past_key_values A__ : Union[str, Any] =seqlen + 2 A__ : List[Any] =( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) A__ : Optional[Any] =[ (torch.zeros(UpperCamelCase__ ), torch.zeros(UpperCamelCase__ )) for _ in range(self.num_layers ) ] A__ : Optional[Any] =common_inputs["attention_mask"] if self.use_past: A__ : Any =ordered_inputs["attention_mask"].dtype A__ : Tuple =torch.cat( [ordered_inputs["attention_mask"], torch.ones(UpperCamelCase__ , UpperCamelCase__ , dtype=UpperCamelCase__ )] , dim=1 ) return ordered_inputs @property def _UpperCAmelCase ( self : List[str] ): return 13
656
0
"""simple docstring""" import math def UpperCamelCase ( SCREAMING_SNAKE_CASE_ = 100 ) ->Dict: _lowerCamelCase : str = sum(i * i for i in range(1 , n + 1 ) ) _lowerCamelCase : Any = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
434
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A : Union[str, Any] = logging.get_logger(__name__) __A : Any = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Tuple = """megatron-bert""" def __init__( self : Tuple , UpperCamelCase__ : Dict=29056 , UpperCamelCase__ : int=1024 , UpperCamelCase__ : Optional[int]=24 , UpperCamelCase__ : Dict=16 , UpperCamelCase__ : int=4096 , UpperCamelCase__ : str="gelu" , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : int=0.1 , UpperCamelCase__ : int=512 , UpperCamelCase__ : str=2 , UpperCamelCase__ : Union[str, Any]=0.02 , UpperCamelCase__ : Any=1E-12 , UpperCamelCase__ : List[Any]=0 , UpperCamelCase__ : str="absolute" , UpperCamelCase__ : Dict=True , **UpperCamelCase__ : Tuple , ): super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) A__ : Optional[int] =vocab_size A__ : Optional[int] =hidden_size A__ : str =num_hidden_layers A__ : Any =num_attention_heads A__ : str =hidden_act A__ : Optional[int] =intermediate_size A__ : str =hidden_dropout_prob A__ : str =attention_probs_dropout_prob A__ : List[Any] =max_position_embeddings A__ : List[Any] =type_vocab_size A__ : Tuple =initializer_range A__ : Any =layer_norm_eps A__ : Any =position_embedding_type A__ : Union[str, Any] =use_cache
656
0
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig __A : Optional[Any] = logging.get_logger(__name__) # General docstring __A : str = "PoolFormerConfig" # Base docstring __A : Optional[Any] = "sail/poolformer_s12" __A : List[Any] = [1, 5_1_2, 7, 7] # Image classification docstring __A : List[str] = "sail/poolformer_s12" __A : Tuple = "tabby, tabby cat" __A : Tuple = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def __a ( A__ : Any , A__ : float = 0.0 , A__ : bool = False ): if drop_prob == 0.0 or not training: return input SCREAMING_SNAKE_CASE = 1 - drop_prob SCREAMING_SNAKE_CASE = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets SCREAMING_SNAKE_CASE = keep_prob + torch.rand(A__ , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize SCREAMING_SNAKE_CASE = input.div(A__ ) * random_tensor return output class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Optional[int] , __lowerCamelCase : Optional[float] = None ): super().__init__() SCREAMING_SNAKE_CASE = drop_prob def _snake_case ( self : List[str] , __lowerCamelCase : torch.Tensor ): return drop_path(UpperCamelCase__ , self.drop_prob , self.training ) def _snake_case ( self : List[str] ): return "p={}".format(self.drop_prob ) class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int=None ): super().__init__() SCREAMING_SNAKE_CASE = patch_size if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (patch_size, patch_size) SCREAMING_SNAKE_CASE = stride if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (stride, stride) SCREAMING_SNAKE_CASE = padding if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (padding, padding) SCREAMING_SNAKE_CASE = nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=UpperCamelCase__ ) SCREAMING_SNAKE_CASE = norm_layer(UpperCamelCase__ ) if norm_layer else nn.Identity() def _snake_case ( self : Tuple , __lowerCamelCase : str ): SCREAMING_SNAKE_CASE = self.projection(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = self.norm(UpperCamelCase__ ) return embeddings class _SCREAMING_SNAKE_CASE ( nn.GroupNorm ): '''simple docstring''' def __init__( self : Tuple , __lowerCamelCase : Dict , **__lowerCamelCase : Union[str, Any] ): super().__init__(1 , UpperCamelCase__ , **UpperCamelCase__ ) class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Tuple , __lowerCamelCase : Optional[int] ): super().__init__() SCREAMING_SNAKE_CASE = nn.AvgPoolad(UpperCamelCase__ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCamelCase__ ) def _snake_case ( self : List[str] , __lowerCamelCase : List[str] ): return self.pool(UpperCamelCase__ ) - hidden_states class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : Dict , __lowerCamelCase : Any , __lowerCamelCase : Union[str, Any] ): super().__init__() SCREAMING_SNAKE_CASE = nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) SCREAMING_SNAKE_CASE = nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) SCREAMING_SNAKE_CASE = PoolFormerDropPath(UpperCamelCase__ ) if isinstance(config.hidden_act , UpperCamelCase__ ): SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: SCREAMING_SNAKE_CASE = config.hidden_act def _snake_case ( self : Tuple , __lowerCamelCase : Dict ): SCREAMING_SNAKE_CASE = self.conva(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = self.act_fn(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = self.drop(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = self.conva(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = self.drop(UpperCamelCase__ ) return hidden_states class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Any ): super().__init__() SCREAMING_SNAKE_CASE = PoolFormerPooling(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = PoolFormerOutput(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(UpperCamelCase__ ) # Useful for training neural nets SCREAMING_SNAKE_CASE = PoolFormerDropPath(UpperCamelCase__ ) if drop_path > 0.0 else nn.Identity() SCREAMING_SNAKE_CASE = config.use_layer_scale if config.use_layer_scale: SCREAMING_SNAKE_CASE = nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) SCREAMING_SNAKE_CASE = nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) def _snake_case ( self : Any , __lowerCamelCase : Optional[int] ): if self.use_layer_scale: SCREAMING_SNAKE_CASE = self.pooling(self.before_norm(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection SCREAMING_SNAKE_CASE = hidden_states + self.drop_path(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = () SCREAMING_SNAKE_CASE = self.output(self.after_norm(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection SCREAMING_SNAKE_CASE = hidden_states + self.drop_path(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = (output,) + outputs return outputs else: SCREAMING_SNAKE_CASE = self.drop_path(self.pooling(self.before_norm(UpperCamelCase__ ) ) ) # First residual connection SCREAMING_SNAKE_CASE = pooling_output + hidden_states SCREAMING_SNAKE_CASE = () # Second residual connection inside the PoolFormerOutput block SCREAMING_SNAKE_CASE = self.drop_path(self.output(self.after_norm(UpperCamelCase__ ) ) ) SCREAMING_SNAKE_CASE = hidden_states + layer_output SCREAMING_SNAKE_CASE = (output,) + outputs return outputs class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Dict , __lowerCamelCase : List[str] ): super().__init__() SCREAMING_SNAKE_CASE = config # stochastic depth decay rule SCREAMING_SNAKE_CASE = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings SCREAMING_SNAKE_CASE = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) SCREAMING_SNAKE_CASE = nn.ModuleList(UpperCamelCase__ ) # Transformer blocks SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers SCREAMING_SNAKE_CASE = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCamelCase__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE = nn.ModuleList(UpperCamelCase__ ) def _snake_case ( self : Dict , __lowerCamelCase : Dict , __lowerCamelCase : Tuple=False , __lowerCamelCase : Optional[int]=True ): SCREAMING_SNAKE_CASE = () if output_hidden_states else None SCREAMING_SNAKE_CASE = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): SCREAMING_SNAKE_CASE = layers # Get patch embeddings from hidden_states SCREAMING_SNAKE_CASE = embedding_layer(UpperCamelCase__ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE = blk(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = layer_outputs[0] if output_hidden_states: SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ ) class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' lowerCamelCase__ = PoolFormerConfig lowerCamelCase__ = """poolformer""" lowerCamelCase__ = """pixel_values""" lowerCamelCase__ = True def _snake_case ( self : List[str] , __lowerCamelCase : str ): if isinstance(UpperCamelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def _snake_case ( self : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any]=False ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE = value __A : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( "The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top." , _UpperCamelCase , ) class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' def __init__( self : List[str] , __lowerCamelCase : Dict ): super().__init__(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = config SCREAMING_SNAKE_CASE = PoolFormerEncoder(UpperCamelCase__ ) # Initialize weights and apply final processing self.post_init() def _snake_case ( self : Tuple ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _snake_case ( self : str , __lowerCamelCase : Optional[torch.FloatTensor] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[bool] = None , ): SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) SCREAMING_SNAKE_CASE = self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__( self : Dict , __lowerCamelCase : Optional[Any] ): super().__init__() SCREAMING_SNAKE_CASE = nn.Linear(config.hidden_size , config.hidden_size ) def _snake_case ( self : Optional[Any] , __lowerCamelCase : List[Any] ): SCREAMING_SNAKE_CASE = self.dense(UpperCamelCase__ ) return output @add_start_docstrings( "\n PoolFormer Model transformer with an image classification head on top\n " , _UpperCamelCase , ) class _SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : str ): super().__init__(UpperCamelCase__ ) SCREAMING_SNAKE_CASE = config.num_labels SCREAMING_SNAKE_CASE = PoolFormerModel(UpperCamelCase__ ) # Final norm SCREAMING_SNAKE_CASE = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head SCREAMING_SNAKE_CASE = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _snake_case ( self : Optional[int] , __lowerCamelCase : Optional[torch.FloatTensor] = None , __lowerCamelCase : Optional[torch.LongTensor] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[bool] = None , ): SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict SCREAMING_SNAKE_CASE = self.poolformer( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE = outputs[0] SCREAMING_SNAKE_CASE = self.classifier(self.norm(UpperCamelCase__ ).mean([-2, -1] ) ) SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: SCREAMING_SNAKE_CASE = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): SCREAMING_SNAKE_CASE = "single_label_classification" else: SCREAMING_SNAKE_CASE = "multi_label_classification" if self.config.problem_type == "regression": SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: SCREAMING_SNAKE_CASE = loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": SCREAMING_SNAKE_CASE = CrossEntropyLoss() SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() SCREAMING_SNAKE_CASE = loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states )
16
"""simple docstring""" from __future__ import annotations def lowercase ( UpperCamelCase : list[float] ): """simple docstring""" if len(UpperCamelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) A__ : Union[str, Any] =nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
656
0
def lowerCamelCase( a__ ,a__ ,a__ ,a__): # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path) def lowerCamelCase( a__ ,a__ ,a__): # Base Case if curr_ind == len(a__): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 ,len(a__)): if valid_connection(a__ ,a__ ,a__ ,a__): # Insert current vertex into path as next transition _SCREAMING_SNAKE_CASE =next_ver # Validate created path if util_hamilton_cycle(a__ ,a__ ,curr_ind + 1): return True # Backtrack _SCREAMING_SNAKE_CASE =-1 return False def lowerCamelCase( a__ ,a__ = 0): _SCREAMING_SNAKE_CASE =[-1] * (len(a__) + 1) # initialize start and end of path with starting index _SCREAMING_SNAKE_CASE =start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(a__ ,a__ ,1) else []
691
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : Optional[Any] = { "configuration_mega": ["MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegaConfig", "MegaOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ "MEGA_PRETRAINED_MODEL_ARCHIVE_LIST", "MegaForCausalLM", "MegaForMaskedLM", "MegaForMultipleChoice", "MegaForQuestionAnswering", "MegaForSequenceClassification", "MegaForTokenClassification", "MegaModel", "MegaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
656
0
from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _lowerCamelCase : int = 0 _lowerCamelCase : List[Any] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _lowerCamelCase : Tuple = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _lowerCamelCase : List[str] = tuple[int, int] class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : int , lowercase : int , lowercase : int , lowercase : int , lowercase : int , lowercase : Node | None , ): '''simple docstring''' _snake_case = pos_x _snake_case = pos_y _snake_case = (pos_y, pos_x) _snake_case = goal_x _snake_case = goal_y _snake_case = g_cost _snake_case = parent _snake_case = self.calculate_heuristic() _snake_case = self.g_cost + self.h_cost def A ( self : List[Any] ): '''simple docstring''' _snake_case = self.pos_x - self.goal_x _snake_case = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(UpperCamelCase__ ) + abs(UpperCamelCase__ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self : List[Any] , lowercase : Node ): '''simple docstring''' return self.f_cost < other.f_cost class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Any , lowercase : TPosition , lowercase : TPosition ): '''simple docstring''' _snake_case = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , UpperCamelCase__ ) _snake_case = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , UpperCamelCase__ ) _snake_case = [self.start] _snake_case = [] _snake_case = False def A ( self : Optional[int] ): '''simple docstring''' while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() _snake_case = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(UpperCamelCase__ ) self.closed_nodes.append(UpperCamelCase__ ) _snake_case = self.get_successors(UpperCamelCase__ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(UpperCamelCase__ ) else: # retrieve the best current path _snake_case = self.open_nodes.pop(self.open_nodes.index(UpperCamelCase__ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(UpperCamelCase__ ) else: self.open_nodes.append(UpperCamelCase__ ) return [self.start.pos] def A ( self : List[str] , lowercase : Node ): '''simple docstring''' _snake_case = [] for action in delta: _snake_case = parent.pos_x + action[1] _snake_case = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(UpperCamelCase__ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( UpperCamelCase__ , UpperCamelCase__ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , UpperCamelCase__ , ) ) return successors def A ( self : str , lowercase : Node | None ): '''simple docstring''' _snake_case = node _snake_case = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) _snake_case = current_node.parent path.reverse() return path class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : TPosition , lowercase : TPosition ): '''simple docstring''' _snake_case = AStar(UpperCamelCase__ , UpperCamelCase__ ) _snake_case = AStar(UpperCamelCase__ , UpperCamelCase__ ) _snake_case = False def A ( self : Dict ): '''simple docstring''' while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() _snake_case = self.fwd_astar.open_nodes.pop(0 ) _snake_case = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( UpperCamelCase__ , UpperCamelCase__ ) self.fwd_astar.closed_nodes.append(UpperCamelCase__ ) self.bwd_astar.closed_nodes.append(UpperCamelCase__ ) _snake_case = current_bwd_node _snake_case = current_fwd_node _snake_case = { self.fwd_astar: self.fwd_astar.get_successors(UpperCamelCase__ ), self.bwd_astar: self.bwd_astar.get_successors(UpperCamelCase__ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(UpperCamelCase__ ) else: # retrieve the best current path _snake_case = astar.open_nodes.pop( astar.open_nodes.index(UpperCamelCase__ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(UpperCamelCase__ ) else: astar.open_nodes.append(UpperCamelCase__ ) return [self.fwd_astar.start.pos] def A ( self : List[str] , lowercase : Node , lowercase : Node ): '''simple docstring''' _snake_case = self.fwd_astar.retrace_path(UpperCamelCase__ ) _snake_case = self.bwd_astar.retrace_path(UpperCamelCase__ ) bwd_path.pop() bwd_path.reverse() _snake_case = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _lowerCamelCase : Union[str, Any] = (0, 0) _lowerCamelCase : int = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _lowerCamelCase : int = time.time() _lowerCamelCase : Any = AStar(init, goal) _lowerCamelCase : List[Any] = a_star.search() _lowerCamelCase : Any = time.time() - start_time print(F'AStar execution time = {end_time:f} seconds') _lowerCamelCase : Union[str, Any] = time.time() _lowerCamelCase : List[Any] = BidirectionalAStar(init, goal) _lowerCamelCase : Optional[int] = time.time() - bd_start_time print(F'BidirectionalAStar execution time = {bd_end_time:f} seconds')
686
"""simple docstring""" def lowercase ( UpperCamelCase : int ): """simple docstring""" if num <= 0: raise ValueError("Input must be a positive integer" ) A__ : Union[str, Any] =[True] * (num + 1) A__ : Union[str, Any] =2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCamelCase ): A__ : str =False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
656
0
"""simple docstring""" import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def snake_case ( self ): __lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) __lowerCAmelCase = "A painting of a squirrel eating a burger" __lowerCAmelCase = jax.device_count() __lowerCAmelCase = num_samples * [prompt] __lowerCAmelCase = sd_pipe.prepare_inputs(UpperCamelCase__ ) __lowerCAmelCase = replicate(UpperCamelCase__ ) __lowerCAmelCase = shard(UpperCamelCase__ ) __lowerCAmelCase = jax.random.PRNGKey(0 ) __lowerCAmelCase = jax.random.split(UpperCamelCase__ , jax.device_count() ) __lowerCAmelCase = sd_pipe(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , num_inference_steps=25 , jit=UpperCamelCase__ )[0] assert images.shape == (jax.device_count(), 1, 7_68, 7_68, 3) __lowerCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __lowerCAmelCase = images[0, 2_53:2_56, 2_53:2_56, -1] __lowerCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) __lowerCAmelCase = jnp.array([0.4_2_3_8, 0.4_4_1_4, 0.4_3_9_5, 0.4_4_5_3, 0.4_6_2_9, 0.4_5_9_0, 0.4_5_3_1, 0.4_5_5_0_8, 0.4_5_1_2] ) print(f"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def snake_case ( self ): __lowerCAmelCase = "stabilityai/stable-diffusion-2" __lowerCAmelCase = FlaxDPMSolverMultistepScheduler.from_pretrained(UpperCamelCase__ , subfolder="scheduler" ) __lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( UpperCamelCase__ , scheduler=UpperCamelCase__ , revision="bf16" , dtype=jnp.bfloataa , ) __lowerCAmelCase = scheduler_params __lowerCAmelCase = "A painting of a squirrel eating a burger" __lowerCAmelCase = jax.device_count() __lowerCAmelCase = num_samples * [prompt] __lowerCAmelCase = sd_pipe.prepare_inputs(UpperCamelCase__ ) __lowerCAmelCase = replicate(UpperCamelCase__ ) __lowerCAmelCase = shard(UpperCamelCase__ ) __lowerCAmelCase = jax.random.PRNGKey(0 ) __lowerCAmelCase = jax.random.split(UpperCamelCase__ , jax.device_count() ) __lowerCAmelCase = sd_pipe(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , num_inference_steps=25 , jit=UpperCamelCase__ )[0] assert images.shape == (jax.device_count(), 1, 7_68, 7_68, 3) __lowerCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __lowerCAmelCase = images[0, 2_53:2_56, 2_53:2_56, -1] __lowerCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) __lowerCAmelCase = jnp.array([0.4_3_3_6, 0.4_2_9_6_9, 0.4_4_5_3, 0.4_1_9_9, 0.4_2_9_7, 0.4_5_3_1, 0.4_4_3_4, 0.4_4_3_4, 0.4_2_9_7] ) print(f"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
636
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def _UpperCAmelCase ( self : List[Any] ): A__ : Tuple =torch.nn.Linear(10 , 10 ) A__ : List[str] =torch.optim.SGD(model.parameters() , 0.1 ) A__ : Union[str, Any] =Accelerator() A__ : str =accelerator.prepare(UpperCamelCase__ ) try: pickle.loads(pickle.dumps(UpperCamelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer _lowercase = logging.get_logger(__name__) _lowercase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} _lowercase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } _lowercase = { "roberta-base": 5_12, "roberta-large": 5_12, "roberta-large-mnli": 5_12, "distilroberta-base": 5_12, "roberta-base-openai-detector": 5_12, "roberta-large-openai-detector": 5_12, } class lowerCAmelCase_ ( _UpperCamelCase ): '''simple docstring''' _lowerCamelCase: Union[str, Any] = VOCAB_FILES_NAMES _lowerCamelCase: str = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase: Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase: List[str] = ["""input_ids""", """attention_mask"""] _lowerCamelCase: Any = RobertaTokenizer def __init__( self : Optional[Any] ,A_ : List[Any]=None ,A_ : List[str]=None ,A_ : Union[str, Any]=None ,A_ : Optional[int]="replace" ,A_ : Any="<s>" ,A_ : List[str]="</s>" ,A_ : Tuple="</s>" ,A_ : str="<s>" ,A_ : int="<unk>" ,A_ : Any="<pad>" ,A_ : Optional[int]="<mask>" ,A_ : Union[str, Any]=False ,A_ : Optional[Any]=True ,**A_ : int ,) -> str: super().__init__( UpperCamelCase__ ,UpperCamelCase__ ,tokenizer_file=UpperCamelCase__ ,errors=UpperCamelCase__ ,bos_token=UpperCamelCase__ ,eos_token=UpperCamelCase__ ,sep_token=UpperCamelCase__ ,cls_token=UpperCamelCase__ ,unk_token=UpperCamelCase__ ,pad_token=UpperCamelCase__ ,mask_token=UpperCamelCase__ ,add_prefix_space=UpperCamelCase__ ,trim_offsets=UpperCamelCase__ ,**UpperCamelCase__ ,) A = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' ,UpperCamelCase__ ) != add_prefix_space: A = getattr(UpperCamelCase__ ,pre_tok_state.pop('type' ) ) A = add_prefix_space A = pre_tok_class(**UpperCamelCase__ ) A = add_prefix_space A = "post_processor" A = getattr(self.backend_tokenizer ,UpperCamelCase__ ,UpperCamelCase__ ) if tokenizer_component_instance: A = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: A = tuple(state['sep'] ) if "cls" in state: A = tuple(state['cls'] ) A = False if state.get('add_prefix_space' ,UpperCamelCase__ ) != add_prefix_space: A = add_prefix_space A = True if state.get('trim_offsets' ,UpperCamelCase__ ) != trim_offsets: A = trim_offsets A = True if changes_to_apply: A = getattr(UpperCamelCase__ ,state.pop('type' ) ) A = component_class(**UpperCamelCase__ ) setattr(self.backend_tokenizer ,UpperCamelCase__ ,UpperCamelCase__ ) @property def _SCREAMING_SNAKE_CASE ( self : Any ) -> Any: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def _SCREAMING_SNAKE_CASE ( self : Any ,A_ : str ) -> Union[str, Any]: A = AddedToken(UpperCamelCase__ ,lstrip=UpperCamelCase__ ,rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ ,UpperCamelCase__ ) else value A = value def _SCREAMING_SNAKE_CASE ( self : Optional[int] ,*A_ : Any ,**A_ : Optional[Any] ) -> Any: A = kwargs.get('is_split_into_words' ,UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase__ ,**UpperCamelCase__ ) def _SCREAMING_SNAKE_CASE ( self : Tuple ,*A_ : Optional[int] ,**A_ : List[Any] ) -> Any: A = kwargs.get('is_split_into_words' ,UpperCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase__ ,**UpperCamelCase__ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ,A_ : str ,A_ : Optional[str] = None ) -> List[Any]: A = self._tokenizer.model.save(UpperCamelCase__ ,name=UpperCamelCase__ ) return tuple(UpperCamelCase__ ) def _SCREAMING_SNAKE_CASE ( self : Tuple ,A_ : List[str] ,A_ : Optional[Any]=None ) -> List[str]: A = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _SCREAMING_SNAKE_CASE ( self : List[str] ,A_ : List[int] ,A_ : Optional[List[int]] = None ) -> str: A = [self.sep_token_id] A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
91
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: __A : Optional[int] = None __A : Union[str, Any] = logging.get_logger(__name__) __A : List[Any] = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __A : str = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } __A : List[str] = { "google/bigbird-roberta-base": 4_096, "google/bigbird-roberta-large": 4_096, "google/bigbird-base-trivia-itc": 4_096, } __A : Tuple = "▁" class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : Dict = VOCAB_FILES_NAMES __magic_name__ : Any = PRETRAINED_VOCAB_FILES_MAP __magic_name__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ : List[Any] = BigBirdTokenizer __magic_name__ : Any = ["""input_ids""", """attention_mask"""] __magic_name__ : List[int] = [] def __init__( self : str , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : int="</s>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[Any]="[SEP]" , UpperCamelCase__ : List[Any]="[MASK]" , UpperCamelCase__ : str="[CLS]" , **UpperCamelCase__ : List[Any] , ): A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else bos_token A__ : Optional[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else eos_token A__ : Optional[int] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else unk_token A__ : int =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else pad_token A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else cls_token A__ : List[Any] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it A__ : str =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) A__ : List[Any] =vocab_file A__ : Optional[int] =False if not self.vocab_file else True def _UpperCAmelCase ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1] def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : Tuple =[self.sep_token_id] A__ : Dict =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : List[str] =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
656
0
"""simple docstring""" def _UpperCamelCase ( A ): UpperCamelCase_ =set() # edges = list of graph's edges UpperCamelCase_ =get_edges(A ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: UpperCamelCase_ =edges.pop() chosen_vertices.add(A ) chosen_vertices.add(A ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(A ) return chosen_vertices def _UpperCamelCase ( A ): UpperCamelCase_ =set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
391
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A : Optional[int] = logging.get_logger(__name__) __A : Optional[int] = {"vocab_file": "spiece.model"} __A : List[Any] = { "vocab_file": { "TsinghuaAI/CPM-Generate": "https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model", } } class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any]=False , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : Dict="<s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : Union[str, Any]="<unk>" , UpperCamelCase__ : Optional[int]="<sep>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[int]="<cls>" , UpperCamelCase__ : List[str]="<mask>" , UpperCamelCase__ : Optional[Any]=["<eop>", "<eod>"] , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : Dict , ): A__ : List[str] =AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token A__ : Tuple ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) A__ : Dict =3 A__ : int =do_lower_case A__ : str =remove_space A__ : Optional[Any] =keep_accents A__ : int =vocab_file A__ : Dict =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) A__ : Union[str, Any] =jieba A__ : List[str] =str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _UpperCAmelCase ( self : Union[str, Any] ): return len(self.sp_model ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Any ={self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[str] ): A__ : Union[str, Any] =self.__dict__.copy() A__ : Tuple =None return state def __setstate__( self : Tuple , UpperCamelCase__ : int ): A__ : Union[str, Any] =d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A__ : Optional[int] ={} A__ : Union[str, Any] =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : Dict ): if self.remove_space: A__ : Optional[int] =" ".join(inputs.strip().split() ) else: A__ : Optional[Any] =inputs A__ : Any =outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: A__ : Optional[Any] =unicodedata.normalize("NFKD" , UpperCamelCase__ ) A__ : Tuple ="".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: A__ : str =outputs.lower() return outputs def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : str ): A__ : Optional[int] =self.preprocess_text(UpperCamelCase__ ) A__ : Dict =self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) A__ : List[str] =[] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): A__ : str =self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : Union[str, Any] =cur_pieces[1:] else: A__ : List[str] =cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def _UpperCAmelCase ( self : int , UpperCamelCase__ : str ): return self.sp_model.PieceToId(UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[Any] ): return self.sp_model.IdToPiece(UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] , UpperCamelCase__ : str ): A__ : Optional[int] ="".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : str =[self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def _UpperCAmelCase ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): A__ : List[str] =[self.sep_token_id] A__ : Optional[Any] =[2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return A__ : Tuple =os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: A__ : Optional[Any] =self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def _UpperCAmelCase ( self : str , *UpperCamelCase__ : Optional[Any] , **UpperCamelCase__ : int ): A__ : List[Any] =super()._decode(*UpperCamelCase__ , **UpperCamelCase__ ) A__ : Union[str, Any] =text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
656
0
'''simple docstring''' import argparse import os from accelerate.test_utils import execute_subprocess_async def a_ ( UpperCamelCase_=None ): if subparsers is not None: A_ = subparsers.add_parser("test" ) else: A_ = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=UpperCamelCase_ , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase_ ) return parser def a_ ( UpperCamelCase_ ): A_ = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: A_ = script_name else: A_ = f"--config_file={args.config_file} {script_name}" A_ = ["accelerate-launch"] + test_args.split() A_ = execute_subprocess_async(UpperCamelCase_ , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def a_ ( ): A_ = test_command_parser() A_ = parser.parse_args() test_command(UpperCamelCase_ ) if __name__ == "__main__": main()
452
"""simple docstring""" def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations(UpperCamelCase : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" def count_of_possible_combinations_with_dp_array( UpperCamelCase : int , UpperCamelCase : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] A__ : str =sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase ) for item in array ) A__ : List[str] =answer return answer A__ : List[Any] =[-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase , UpperCamelCase ) def lowercase ( UpperCamelCase : int , UpperCamelCase : list[int] , UpperCamelCase : int ): """simple docstring""" A__ : str =[0] * (target + 1) A__ : Optional[Any] =1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[Any] = 3 __A : Optional[Any] = 5 __A : int = [1, 2, 5] print(combination_sum_iv(n, array, target))
656
0
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class __snake_case : def __init__( self: int , A_: Dict , A_: int = 13 , A_: int = 64 , A_: int = 2 , A_: int = 3 , A_: int = 3 , A_: bool = True , A_: bool = True , A_: int = 1_28 , A_: Optional[int]=[16, 32, 64, 1_28] , A_: int = 7 , A_: int = 4 , A_: int = 37 , A_: str = "gelu" , A_: float = 0.1 , A_: float = 0.1 , A_: int = 10 , A_: float = 0.02 , A_: int = 2 , A_: int = 1 , A_: int = 1_28 , A_: List[int] = [2, 2, 2, 2] , A_: int = 2 , A_: int = 2 , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = encoder_stride __lowerCamelCase = num_attention_outputs __lowerCamelCase = embed_dim __lowerCamelCase = embed_dim + 1 __lowerCamelCase = resolution __lowerCamelCase = depths __lowerCamelCase = hidden_sizes __lowerCamelCase = dim __lowerCamelCase = mlp_expansion_ratio def __a ( self: str ): __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def __a ( self: List[Any] ): return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def __a ( self: Optional[int] , A_: Optional[Any] , A_: Any , A_: Any ): __lowerCamelCase = TFEfficientFormerModel(config=UpperCamelCase__ ) __lowerCamelCase = model(UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self: Dict , A_: List[Any] , A_: Optional[Any] , A_: Tuple ): __lowerCamelCase = self.type_sequence_label_size __lowerCamelCase = TFEfficientFormerForImageClassification(UpperCamelCase__ ) __lowerCamelCase = model(UpperCamelCase__ , labels=UpperCamelCase__ , training=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCamelCase = 1 __lowerCamelCase = TFEfficientFormerForImageClassification(UpperCamelCase__ ) __lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __a ( self: Optional[int] ): __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase = config_and_inputs __lowerCamelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class __snake_case (_UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __a = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) __a = ( { """feature-extraction""": TFEfficientFormerModel, """image-classification""": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) __a = False __a = False __a = False __a = False __a = False def __a ( self: Tuple ): __lowerCamelCase = TFEfficientFormerModelTester(self ) __lowerCamelCase = ConfigTester( self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def __a ( self: List[Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def __a ( self: str ): pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def __a ( self: Tuple ): pass def __a ( self: List[str] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(UpperCamelCase__ ) __lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def __a ( self: Optional[Any] ): def check_hidden_states_output(A_: List[Any] , A_: Dict , A_: int ): __lowerCamelCase = model_class(UpperCamelCase__ ) __lowerCamelCase = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) , training=UpperCamelCase__ ) __lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowerCamelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) if hasattr(self.model_tester , """encoder_seq_length""" ): __lowerCamelCase = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __lowerCamelCase = seq_length * self.model_tester.chunk_length else: __lowerCamelCase = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __lowerCamelCase = outputs.decoder_hidden_states self.asseretIsInstance(UpperCamelCase__ , (list, tuple) ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) __lowerCamelCase = getattr(self.model_tester , """seq_length""" , UpperCamelCase__ ) __lowerCamelCase = getattr(self.model_tester , """decoder_seq_length""" , UpperCamelCase__ ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def __a ( self: List[Any] , A_: Dict , A_: List[Any] , A_: Optional[int]=False ): __lowerCamelCase = super()._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __a ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def __a ( self: List[str] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCamelCase__ ) def __a ( self: List[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def __a ( self: Any ): for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = TFEfficientFormerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def __a ( self: Any ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = True __lowerCamelCase = getattr(self.model_tester , """seq_length""" , UpperCamelCase__ ) __lowerCamelCase = getattr(self.model_tester , """encoder_seq_length""" , UpperCamelCase__ ) __lowerCamelCase = getattr(self.model_tester , """key_length""" , UpperCamelCase__ ) __lowerCamelCase = getattr(self.model_tester , """chunk_length""" , UpperCamelCase__ ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __lowerCamelCase = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __lowerCamelCase = True __lowerCamelCase = False __lowerCamelCase = True __lowerCamelCase = model_class(UpperCamelCase__ ) __lowerCamelCase = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) , training=UpperCamelCase__ ) __lowerCamelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowerCamelCase = True __lowerCamelCase = model_class(UpperCamelCase__ ) __lowerCamelCase = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) , training=UpperCamelCase__ ) __lowerCamelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def __a ( self: Tuple ): # We use a simplified version of this test for EfficientFormer because it requires training=False # and Keras refuses to let us force that during functional construction __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __lowerCamelCase = model_class(UpperCamelCase__ ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __lowerCamelCase = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=UpperCamelCase__ ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __lowerCamelCase = model(UpperCamelCase__ ) self.assertTrue(outputs_dict is not None ) def a_ ( ): __lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __snake_case (unittest.TestCase ): @cached_property def __a ( self: int ): return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def __a ( self: Dict ): __lowerCamelCase = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # forward pass __lowerCamelCase = model(**UpperCamelCase__ , training=UpperCamelCase__ ) # verify the logits __lowerCamelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) __lowerCamelCase = tf.constant([-0.0_555, 0.4_825, -0.0_852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def __a ( self: List[Any] ): __lowerCamelCase = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=UpperCamelCase__ , return_tensors="""tf""" ) # forward pass __lowerCamelCase = model(**UpperCamelCase__ , training=UpperCamelCase__ ) # verify the logits __lowerCamelCase = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) __lowerCamelCase = tf.constant([-0.1_312, 0.4_353, -1.0_499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1E-4 ) )
281
"""simple docstring""" import math import tensorflow as tf from packaging import version def lowercase ( UpperCamelCase : Optional[Any] ): """simple docstring""" A__ : List[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : Optional[Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : Tuple =tf.cast(math.pi , x.dtype ) A__ : Dict =tf.cast(0.04_47_15 , x.dtype ) A__ : Optional[int] =0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(UpperCamelCase , 3 )) )) return x * cdf def lowercase ( UpperCamelCase : Optional[int] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) return x * tf.tanh(tf.math.softplus(UpperCamelCase ) ) def lowercase ( UpperCamelCase : List[str] ): """simple docstring""" A__ : Union[str, Any] =tf.convert_to_tensor(UpperCamelCase ) A__ : List[Any] =tf.cast(0.04_47_15 , x.dtype ) A__ : List[Any] =tf.cast(0.79_78_84_56_08 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" A__ : List[str] =tf.convert_to_tensor(UpperCamelCase ) A__ : str =tf.cast(1.7_02 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" return tf.clip_by_value(_gelu(UpperCamelCase ) , -10 , 10 ) def lowercase ( UpperCamelCase : str , UpperCamelCase : Any=-1 ): """simple docstring""" A__ , A__ : Optional[Any] =tf.split(UpperCamelCase , 2 , axis=UpperCamelCase ) return a * tf.math.sigmoid(UpperCamelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def lowercase ( UpperCamelCase : int ): """simple docstring""" return tf.keras.activations.gelu(UpperCamelCase , approximate=UpperCamelCase ) __A : Optional[Any] = tf.keras.activations.gelu __A : Optional[Any] = approximate_gelu_wrap else: __A : Any = _gelu __A : Union[str, Any] = _gelu_new __A : List[str] = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def lowercase ( UpperCamelCase : List[Any] ): """simple docstring""" if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F'''function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}''' )
656
0
import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) _UpperCamelCase : List[Any] =models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation='relu') ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation='relu')) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation='relu')) classifier.add(layers.Dense(units=1, activation='sigmoid')) # Compiling the CNN classifier.compile( optimizer='adam', loss='binary_crossentropy', metrics=['accuracy'] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') _UpperCamelCase : Optional[int] =tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) _UpperCamelCase : Dict =tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) _UpperCamelCase : Tuple =train_datagen.flow_from_directory( 'dataset/training_set', target_size=(64, 64), batch_size=32, class_mode='binary' ) _UpperCamelCase : Any =test_datagen.flow_from_directory( 'dataset/test_set', target_size=(64, 64), batch_size=32, class_mode='binary' ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save('cnn.h5') # Part 3 - Making new predictions _UpperCamelCase : Optional[int] =tf.keras.preprocessing.image.load_img( 'dataset/single_prediction/image.png', target_size=(64, 64) ) _UpperCamelCase : str =tf.keras.preprocessing.image.img_to_array(test_image) _UpperCamelCase : Optional[int] =np.expand_dims(test_image, axis=0) _UpperCamelCase : Union[str, Any] =classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: _UpperCamelCase : Tuple ="Normal" if result[0][0] == 1: _UpperCamelCase : Union[str, Any] ="Abnormality detected"
206
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def _UpperCAmelCase ( self : Dict ): A__ : Optional[Any] =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_attention_heads" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_encoder_blocks" ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any]=13 , UpperCamelCase__ : Tuple=64 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Union[str, Any]=4 , UpperCamelCase__ : Dict=[2, 2, 2, 2] , UpperCamelCase__ : Union[str, Any]=[8, 4, 2, 1] , UpperCamelCase__ : Tuple=[16, 32, 64, 128] , UpperCamelCase__ : Optional[int]=[1, 4, 8, 16] , UpperCamelCase__ : Any=[1, 2, 4, 8] , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Dict="gelu" , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[Any]=None , ): A__ : Tuple =parent A__ : List[Any] =batch_size A__ : List[Any] =image_size A__ : Union[str, Any] =num_channels A__ : Optional[int] =num_encoder_blocks A__ : Any =sr_ratios A__ : Any =depths A__ : List[Any] =hidden_sizes A__ : List[Any] =downsampling_rates A__ : List[str] =num_attention_heads A__ : int =is_training A__ : List[Any] =use_labels A__ : Any =hidden_act A__ : Dict =hidden_dropout_prob A__ : int =attention_probs_dropout_prob A__ : List[Any] =initializer_range A__ : Tuple =num_labels A__ : List[Any] =scope def _UpperCAmelCase ( self : Optional[int] ): A__ : List[str] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Any =None if self.use_labels: A__ : Tuple =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A__ : List[Any] =self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self : Tuple ): return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int ): A__ : Any =SegformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Dict =model(UpperCamelCase__ ) A__ : Optional[int] =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[Any] ): A__ : str =self.num_labels A__ : Optional[Any] =SegformerForSemanticSegmentation(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : Optional[Any] =model(UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) A__ : List[Any] =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str ): A__ : Tuple =1 A__ : Tuple =SegformerForSemanticSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A__ : List[str] =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(UpperCamelCase__ ) A__ : Dict =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertGreater(result.loss , 0.0 ) def _UpperCAmelCase ( self : str ): A__ : Union[str, Any] =self.prepare_config_and_inputs() A__ , A__ , A__ : Tuple =config_and_inputs A__ : Tuple ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Dict = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) __magic_name__ : Optional[int] = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ : Dict = True __magic_name__ : List[str] = False __magic_name__ : Optional[Any] = False __magic_name__ : str = False def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =SegformerModelTester(self ) A__ : Tuple =SegformerConfigTester(self , config_class=UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): self.config_tester.run_common_tests() def _UpperCAmelCase ( self : Dict ): A__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Tuple ): A__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*UpperCamelCase__ ) def _UpperCAmelCase ( self : Union[str, Any] ): A__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*UpperCamelCase__ ) @unittest.skip("SegFormer does not use inputs_embeds" ) def _UpperCAmelCase ( self : Dict ): pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods" ) def _UpperCAmelCase ( self : Tuple ): pass def _UpperCAmelCase ( self : List[str] ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : int =model_class(UpperCamelCase__ ) A__ : Optional[int] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Optional[int] =[*signature.parameters.keys()] A__ : List[str] =["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def _UpperCAmelCase ( self : str ): A__ , A__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() A__ : Union[str, Any] =True for model_class in self.all_model_classes: A__ : Optional[Any] =True A__ : Union[str, Any] =False A__ : str =True A__ : Optional[int] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : str =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Any =outputs.attentions A__ : List[str] =sum(self.model_tester.depths ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ : Dict =True A__ : str =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Any =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Union[str, Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : List[Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) A__ : Tuple =(self.model_tester.image_size // 32) ** 2 A__ : Optional[Any] =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) A__ : int =len(UpperCamelCase__ ) # Check attention is always last and order is fine A__ : Optional[Any] =True A__ : Any =True A__ : Union[str, Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : Optional[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(out_len + 1 , len(UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) A__ : Union[str, Any] =(self.model_tester.image_size // 4) ** 2 A__ : Tuple =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _UpperCAmelCase ( self : List[Any] ): def check_hidden_states_output(UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple ): A__ : Optional[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): A__ : List[Any] =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) A__ : Optional[Any] =outputs.hidden_states A__ : int =self.model_tester.num_encoder_blocks self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Optional[Any] =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : str =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def _UpperCAmelCase ( self : Optional[int] ): if not self.model_tester.is_training: return A__ , A__ : int =self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] =True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ): continue A__ : List[Any] =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() A__ : int =self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) A__ : Union[str, Any] =model(**UpperCamelCase__ ).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _UpperCAmelCase ( self : Tuple ): pass @slow def _UpperCAmelCase ( self : Tuple ): for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Tuple =SegformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowercase ( ): """simple docstring""" A__ : List[Any] =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): # only resize + normalize A__ : List[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : Union[str, Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : Union[str, Any] =prepare_img() A__ : Union[str, Any] =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : int =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : Dict =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : Optional[int] =torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : Union[str, Any] ): # only resize + normalize A__ : Dict =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : int =SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024" ).to(UpperCamelCase__ ) A__ : Tuple =prepare_img() A__ : str =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Optional[int] =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : int =model(UpperCamelCase__ ) A__ : List[str] =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) A__ : List[Any] =torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-1 ) ) @slow def _UpperCAmelCase ( self : int ): # only resize + normalize A__ : Optional[Any] =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) A__ : List[Any] =SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( UpperCamelCase__ ) A__ : str =prepare_img() A__ : Dict =image_processor(images=UpperCamelCase__ , return_tensors="pt" ) A__ : Any =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): A__ : Dict =model(UpperCamelCase__ ) A__ : Any =outputs.logits.detach().cpu() A__ : Union[str, Any] =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ , target_sizes=[(500, 300)] ) A__ : List[str] =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ ) A__ : int =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ ) A__ : Tuple =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ )
656
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { "transfo-xl-wt103": "https://huggingface.co/transfo-xl-wt103/resolve/main/config.json", } class __magic_name__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : int = """transfo-xl""" _SCREAMING_SNAKE_CASE : str = ["""mems"""] _SCREAMING_SNAKE_CASE : int = { """n_token""": """vocab_size""", """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : str , snake_case_ : str=267735 , snake_case_ : Dict=[20000, 40000, 200000] , snake_case_ : Union[str, Any]=1024 , snake_case_ : Union[str, Any]=1024 , snake_case_ : Tuple=16 , snake_case_ : List[str]=64 , snake_case_ : List[str]=4096 , snake_case_ : Dict=4 , snake_case_ : Any=False , snake_case_ : Union[str, Any]=18 , snake_case_ : Any=1600 , snake_case_ : Dict=1000 , snake_case_ : List[str]=True , snake_case_ : str=True , snake_case_ : str=0 , snake_case_ : List[str]=-1 , snake_case_ : List[Any]=True , snake_case_ : Optional[int]=0.1 , snake_case_ : List[Any]=0.0 , snake_case_ : str=True , snake_case_ : Optional[int]="normal" , snake_case_ : int=0.01 , snake_case_ : Tuple=0.01 , snake_case_ : int=0.02 , snake_case_ : List[str]=1e-5 , snake_case_ : Union[str, Any]=0 , **snake_case_ : Tuple , ): __snake_case = vocab_size __snake_case = [] self.cutoffs.extend(UpperCamelCase__ ) if proj_share_all_but_first: __snake_case = [False] + [True] * len(self.cutoffs ) else: __snake_case = [False] + [False] * len(self.cutoffs ) __snake_case = d_model __snake_case = d_embed __snake_case = d_head __snake_case = d_inner __snake_case = div_val __snake_case = pre_lnorm __snake_case = n_layer __snake_case = n_head __snake_case = mem_len __snake_case = same_length __snake_case = attn_type __snake_case = clamp_len __snake_case = sample_softmax __snake_case = adaptive __snake_case = dropout __snake_case = dropatt __snake_case = untie_r __snake_case = init __snake_case = init_range __snake_case = proj_init_std __snake_case = init_std __snake_case = layer_norm_epsilon super().__init__(eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) @property def lowerCAmelCase ( self : Union[str, Any] ): # Message copied from Transformer-XL documentation logger.info(F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def lowerCAmelCase ( self : Optional[int] , snake_case_ : str ): # Message copied from Transformer-XL documentation raise NotImplementedError( F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
163
"""simple docstring""" import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any]=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[str]=99 , UpperCamelCase__ : Optional[Any]=32 , UpperCamelCase__ : Any=5 , UpperCamelCase__ : Dict=4 , UpperCamelCase__ : Union[str, Any]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Any=0.1 , UpperCamelCase__ : Optional[Any]=0.1 , UpperCamelCase__ : Optional[Any]=512 , UpperCamelCase__ : Union[str, Any]=16 , UpperCamelCase__ : List[str]=2 , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : List[Any]=4 , ): A__ : str =parent A__ : List[str] =batch_size A__ : Any =seq_length A__ : List[str] =is_training A__ : List[Any] =use_attention_mask A__ : List[Any] =use_token_type_ids A__ : Dict =use_labels A__ : List[Any] =vocab_size A__ : Optional[int] =hidden_size A__ : Optional[Any] =num_hidden_layers A__ : str =num_attention_heads A__ : int =intermediate_size A__ : Tuple =hidden_act A__ : Tuple =hidden_dropout_prob A__ : Dict =attention_probs_dropout_prob A__ : Any =max_position_embeddings A__ : Any =type_vocab_size A__ : Union[str, Any] =type_sequence_label_size A__ : Optional[Any] =initializer_range A__ : int =num_choices def _UpperCAmelCase ( self : Tuple ): A__ : List[str] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[str] =None if self.use_attention_mask: A__ : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) A__ : str =None if self.use_token_type_ids: A__ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Any =RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCAmelCase ( self : Tuple ): A__ : Dict =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : str =config_and_inputs A__ : Optional[Any] ={"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _UpperCAmelCase ( self : int ): A__ : str =self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Union[str, Any] =config_and_inputs A__ : Union[str, Any] =True A__ : List[Any] =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : Tuple =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class __lowerCAmelCase ( _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : Union[str, Any] = True __magic_name__ : Dict = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCAmelCase ( self : Optional[int] ): A__ : Optional[int] =FlaxRobertaPreLayerNormModelTester(self ) @slow def _UpperCAmelCase ( self : List[Any] ): for model_class_name in self.all_model_classes: A__ : Tuple =model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Union[str, Any] =model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase__ ) @require_flax class __lowerCAmelCase ( unittest.TestCase): '''simple docstring''' @slow def _UpperCAmelCase ( self : Tuple ): A__ : Any =FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : Tuple =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : str =model(UpperCamelCase__ )[0] A__ : List[Any] =[1, 11, 50265] self.assertEqual(list(output.shape ) , UpperCamelCase__ ) # compare the actual values for a slice. A__ : Any =np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self : List[Any] ): A__ : Union[str, Any] =FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=UpperCamelCase__ ) A__ : List[Any] =np.array([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] , dtype=jnp.intaa ) A__ : Dict =model(UpperCamelCase__ )[0] # compare the actual values for a slice. A__ : Optional[Any] =np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , UpperCamelCase__ , atol=1E-4 ) )
656
0
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() SCREAMING_SNAKE_CASE__ : List[Any] =logging.get_logger('transformers.models.speecht5') SCREAMING_SNAKE_CASE__ : Optional[Any] ={ "speech_encoder_prenet.layer_norm": "speecht5.encoder.prenet.feature_projection.layer_norm", "speech_encoder_prenet.post_extract_proj": "speecht5.encoder.prenet.feature_projection.projection", "speech_encoder_prenet.pos_conv.0": "speecht5.encoder.prenet.pos_conv_embed.conv", "speech_encoder_prenet.mask_emb": "speecht5.encoder.prenet.masked_spec_embed", } SCREAMING_SNAKE_CASE__ : Optional[int] ={ "text_encoder_prenet.encoder_prenet.0": "speecht5.encoder.prenet.embed_tokens", "text_encoder_prenet.encoder_prenet.1.alpha": "speecht5.encoder.prenet.encode_positions.alpha", } SCREAMING_SNAKE_CASE__ : List[str] ={ "speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0": "speecht5.decoder.prenet.layers.0", "speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0": "speecht5.decoder.prenet.layers.1", "speech_decoder_prenet.decoder_prenet.0.1": "speecht5.decoder.prenet.final_layer", "speech_decoder_prenet.decoder_prenet.1.alpha": "speecht5.decoder.prenet.encode_positions.alpha", "speech_decoder_prenet.spkembs_layer.0": "speecht5.decoder.prenet.speaker_embeds_layer", } SCREAMING_SNAKE_CASE__ : List[Any] ={ "speech_decoder_postnet.feat_out": "speech_decoder_postnet.feat_out", "speech_decoder_postnet.prob_out": "speech_decoder_postnet.prob_out", "speech_decoder_postnet.postnet.postnet.0.0": "speech_decoder_postnet.layers.0.conv", "speech_decoder_postnet.postnet.postnet.0.1": "speech_decoder_postnet.layers.0.batch_norm", "speech_decoder_postnet.postnet.postnet.1.0": "speech_decoder_postnet.layers.1.conv", "speech_decoder_postnet.postnet.postnet.1.1": "speech_decoder_postnet.layers.1.batch_norm", "speech_decoder_postnet.postnet.postnet.2.0": "speech_decoder_postnet.layers.2.conv", "speech_decoder_postnet.postnet.postnet.2.1": "speech_decoder_postnet.layers.2.batch_norm", "speech_decoder_postnet.postnet.postnet.3.0": "speech_decoder_postnet.layers.3.conv", "speech_decoder_postnet.postnet.postnet.3.1": "speech_decoder_postnet.layers.3.batch_norm", "speech_decoder_postnet.postnet.postnet.4.0": "speech_decoder_postnet.layers.4.conv", "speech_decoder_postnet.postnet.postnet.4.1": "speech_decoder_postnet.layers.4.batch_norm", } SCREAMING_SNAKE_CASE__ : Union[str, Any] ={ "text_decoder_prenet.embed_tokens": "speecht5.decoder.prenet.embed_tokens", } SCREAMING_SNAKE_CASE__ : Any ={ "text_decoder_postnet.output_projection": "text_decoder_postnet.lm_head", } SCREAMING_SNAKE_CASE__ : Union[str, Any] ={ "encoder.layers.*.self_attn.k_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj", "encoder.layers.*.self_attn.v_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj", "encoder.layers.*.self_attn.q_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj", "encoder.layers.*.self_attn.out_proj": "speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj", "encoder.layers.*.self_attn_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.layer_norm", "encoder.layers.*.fc1": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense", "encoder.layers.*.fc2": "speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense", "encoder.layers.*.final_layer_norm": "speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm", "encoder.layer_norm": "speecht5.encoder.wrapped_encoder.layer_norm", "encoder.pos_emb.pe_k": "speecht5.encoder.wrapped_encoder.embed_positions.pe_k", } SCREAMING_SNAKE_CASE__ : Optional[int] ={ "decoder.layers.*.self_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj", "decoder.layers.*.self_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj", "decoder.layers.*.self_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj", "decoder.layers.*.self_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj", "decoder.layers.*.self_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm", "decoder.layers.*.encoder_attn.k_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj", "decoder.layers.*.encoder_attn.v_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj", "decoder.layers.*.encoder_attn.q_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj", "decoder.layers.*.encoder_attn.out_proj": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj", "decoder.layers.*.encoder_attn_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm", "decoder.layers.*.fc1": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense", "decoder.layers.*.fc2": "speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense", "decoder.layers.*.final_layer_norm": "speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm", } SCREAMING_SNAKE_CASE__ : Union[str, Any] ={ **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } SCREAMING_SNAKE_CASE__ : Optional[Any] ={ **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } SCREAMING_SNAKE_CASE__ : Optional[int] ={ **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } SCREAMING_SNAKE_CASE__ : int =[] SCREAMING_SNAKE_CASE__ : int =[ "encoder.version", "encoder.layers.*.norm_k.weight", "encoder.layers.*.norm_k.bias", "decoder.version", "decoder.layers.*.norm_k.weight", "decoder.layers.*.norm_k.bias", "decoder.pos_emb.pe_k", "speech_encoder_prenet.embed_positions._float_tensor", "text_decoder_prenet.embed_positions._float_tensor", ] SCREAMING_SNAKE_CASE__ : Optional[Any] =IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "speech_decoder_prenet.*", "speech_decoder_postnet.*", ] SCREAMING_SNAKE_CASE__ : Tuple =IGNORE_KEYS + [ "encoder.proj", "speech_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] SCREAMING_SNAKE_CASE__ : Union[str, Any] =IGNORE_KEYS + [ "encoder.proj", "text_encoder_prenet.*", "text_decoder_prenet.*", "text_decoder_postnet.*", ] def UpperCamelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ->Optional[Any]: for attribute in key.split('''.''' ): _lowerCamelCase : Dict = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if weight_type is not None: _lowerCamelCase : Union[str, Any] = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ).shape else: _lowerCamelCase : Tuple = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": _lowerCamelCase : Any = value elif weight_type == "weight_g": _lowerCamelCase : Any = value elif weight_type == "weight_v": _lowerCamelCase : Any = value elif weight_type == "bias": _lowerCamelCase : Tuple = value elif weight_type == "running_mean": _lowerCamelCase : Dict = value elif weight_type == "running_var": _lowerCamelCase : List[str] = value elif weight_type == "num_batches_tracked": _lowerCamelCase : Dict = value else: _lowerCamelCase : Optional[int] = value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''' ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ->str: for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: _lowerCamelCase : List[str] = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def UpperCamelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ->str: _lowerCamelCase : Tuple = [] if task == "s2t": _lowerCamelCase : Dict = hf_model.speechta.encoder.prenet.feature_encoder _lowerCamelCase : int = MAPPING_S2T _lowerCamelCase : List[Any] = IGNORE_KEYS_S2T elif task == "t2s": _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : List[Any] = MAPPING_T2S _lowerCamelCase : Tuple = IGNORE_KEYS_T2S elif task == "s2s": _lowerCamelCase : Optional[Any] = hf_model.speechta.encoder.prenet.feature_encoder _lowerCamelCase : Tuple = MAPPING_S2S _lowerCamelCase : Any = IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''' ) for name, value in fairseq_dict.items(): if should_ignore(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): logger.info(F'''{name} was ignored''' ) continue _lowerCamelCase : Optional[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase : List[Any] = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: _lowerCamelCase : Dict = key.split('''.*.''' ) if prefix in name and suffix in name: _lowerCamelCase : int = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: _lowerCamelCase : List[Any] = True if "*" in mapped_key: _lowerCamelCase : Optional[int] = name.split(SCREAMING_SNAKE_CASE_ )[0].split('''.''' )[-2] _lowerCamelCase : int = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE_ ) if "weight_g" in name: _lowerCamelCase : str = "weight_g" elif "weight_v" in name: _lowerCamelCase : Optional[Any] = "weight_v" elif "bias" in name: _lowerCamelCase : Any = "bias" elif "weight" in name: _lowerCamelCase : Optional[int] = "weight" elif "running_mean" in name: _lowerCamelCase : Tuple = "running_mean" elif "running_var" in name: _lowerCamelCase : Optional[int] = "running_var" elif "num_batches_tracked" in name: _lowerCamelCase : str = "num_batches_tracked" else: _lowerCamelCase : List[Any] = None set_recursively(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE_ ) logger.warning(F'''Unused weights: {unused_weights}''' ) def UpperCamelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ->Tuple: _lowerCamelCase : Any = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase : Dict = name.split('''.''' ) _lowerCamelCase : int = int(items[0] ) _lowerCamelCase : str = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) _lowerCamelCase : Optional[Any] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) _lowerCamelCase : Optional[int] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''' ) _lowerCamelCase : Any = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''' ) _lowerCamelCase : Any = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(SCREAMING_SNAKE_CASE_ ) @torch.no_grad() def UpperCamelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , ) ->Any: if config_path is not None: _lowerCamelCase : Any = SpeechTaConfig.from_pretrained(SCREAMING_SNAKE_CASE_ ) else: _lowerCamelCase : Any = SpeechTaConfig() if task == "s2t": _lowerCamelCase : Union[str, Any] = config.max_text_positions _lowerCamelCase : Dict = SpeechTaForSpeechToText(SCREAMING_SNAKE_CASE_ ) elif task == "t2s": _lowerCamelCase : str = 1876 _lowerCamelCase : Optional[int] = 600 _lowerCamelCase : Tuple = config.max_speech_positions _lowerCamelCase : Optional[Any] = SpeechTaForTextToSpeech(SCREAMING_SNAKE_CASE_ ) elif task == "s2s": _lowerCamelCase : str = 1876 _lowerCamelCase : Tuple = config.max_speech_positions _lowerCamelCase : Any = SpeechTaForSpeechToSpeech(SCREAMING_SNAKE_CASE_ ) else: raise ValueError(F'''Unknown task name: {task}''' ) if vocab_path: _lowerCamelCase : str = SpeechTaTokenizer(SCREAMING_SNAKE_CASE_ , model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it _lowerCamelCase : Optional[Any] = AddedToken('''<mask>''' , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) _lowerCamelCase : int = mask_token tokenizer.add_special_tokens({'''mask_token''': mask_token} ) tokenizer.add_tokens(['''<ctc_blank>'''] ) _lowerCamelCase : Dict = SpeechTaFeatureExtractor() _lowerCamelCase : Tuple = SpeechTaProcessor(tokenizer=SCREAMING_SNAKE_CASE_ , feature_extractor=SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) _lowerCamelCase : Union[str, Any] = torch.load(SCREAMING_SNAKE_CASE_ ) recursively_load_weights(fairseq_checkpoint['''model'''] , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) if repo_id: print('''Pushing to the hub...''' ) processor.push_to_hub(SCREAMING_SNAKE_CASE_ ) model.push_to_hub(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Dict =argparse.ArgumentParser() parser.add_argument( '--task', default='s2t', type=str, help='Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--vocab_path', default=None, type=str, help='Path to SentencePiece model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) SCREAMING_SNAKE_CASE__ : str =parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
434
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __A : List[Any] = logging.get_logger(__name__) __A : Any = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] __A : Optional[int] = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def lowercase ( UpperCamelCase : Tuple ): """simple docstring""" A__ : Union[str, Any] =torch.load(UpperCamelCase , map_location="cpu" ) return sd def lowercase ( UpperCamelCase : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : int=rename_keys_prefix ): """simple docstring""" A__ : List[str] =OrderedDict() A__ : str =torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue A__ : Optional[Any] =key for name_pair in rename_keys_prefix: A__ : int =new_key.replace(name_pair[0] , name_pair[1] ) A__ : Dict =d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately A__ : Optional[int] =new_d["cls.predictions.bias"] return new_d @torch.no_grad() def lowercase ( UpperCamelCase : Dict , UpperCamelCase : List[str] ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: A__ : Any ="pretraining" if "vcr" in checkpoint_path: A__ : Union[str, Any] ={"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: A__ : Optional[Any] ={"visual_embedding_dim": 2048} elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048} elif "nlvr" in checkpoint_path: A__ : List[str] ={"visual_embedding_dim": 1024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 512} A__ : List[str] ="multichoice" elif "vqa_advanced" in checkpoint_path: A__ : Any ={"visual_embedding_dim": 2048} A__ : str ="vqa_advanced" elif "vqa" in checkpoint_path: A__ : Optional[int] ={"visual_embedding_dim": 2048, "num_labels": 3129} A__ : str ="vqa" elif "nlvr" in checkpoint_path: A__ : str ={ "visual_embedding_dim": 1024, "num_labels": 2, } A__ : Dict ="nlvr" A__ : Union[str, Any] =VisualBertConfig(**UpperCamelCase ) # Load State Dict A__ : int =load_state_dict(UpperCamelCase ) A__ : Tuple =get_new_dict(UpperCamelCase , UpperCamelCase ) if model_type == "pretraining": A__ : str =VisualBertForPreTraining(UpperCamelCase ) elif model_type == "vqa": A__ : Optional[int] =VisualBertForQuestionAnswering(UpperCamelCase ) elif model_type == "nlvr": A__ : Union[str, Any] =VisualBertForVisualReasoning(UpperCamelCase ) elif model_type == "multichoice": A__ : Union[str, Any] =VisualBertForMultipleChoice(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) # Save Checkpoints Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": __A : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("orig_checkpoint_path", type=str, help="A path to .th on local filesystem.") parser.add_argument("pytorch_dump_folder_path", type=str, help="Path to the output PyTorch model.") __A : str = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
656
0
def __a ( A__ : int , A__ : int ): if b == 0: return 1 if (b % 2) == 0: return actual_power(A__ , int(b / 2 ) ) * actual_power(A__ , int(b / 2 ) ) else: return a * actual_power(A__ , int(b / 2 ) ) * actual_power(A__ , int(b / 2 ) ) def __a ( A__ : int , A__ : int ): if b < 0: return 1 / actual_power(A__ , A__ ) return actual_power(A__ , A__ ) if __name__ == "__main__": print(power(-2, -3))
16
"""simple docstring""" __A : Union[str, Any] = {str(digit): digit**5 for digit in range(10)} def lowercase ( UpperCamelCase : int ): """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase ) ) def lowercase ( ): """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(UpperCamelCase ) ) if __name__ == "__main__": print(solution())
656
0
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class A__ ( _UpperCamelCase ): def __UpperCamelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , '''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , '''num_attention_heads''' ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , '''num_encoder_blocks''' ) ) class A__ : def __init__( self : int , _a : Union[str, Any] , _a : Union[str, Any]=13 , _a : Tuple=64 , _a : Optional[int]=3 , _a : Union[str, Any]=4 , _a : Dict=[2, 2, 2, 2] , _a : Union[str, Any]=[8, 4, 2, 1] , _a : Tuple=[16, 32, 64, 128] , _a : Optional[int]=[1, 4, 8, 16] , _a : Any=[1, 2, 4, 8] , _a : Union[str, Any]=True , _a : str=True , _a : Dict="gelu" , _a : str=0.1 , _a : List[Any]=0.1 , _a : List[str]=0.02 , _a : int=3 , _a : Optional[Any]=None , ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =parent _SCREAMING_SNAKE_CASE =batch_size _SCREAMING_SNAKE_CASE =image_size _SCREAMING_SNAKE_CASE =num_channels _SCREAMING_SNAKE_CASE =num_encoder_blocks _SCREAMING_SNAKE_CASE =sr_ratios _SCREAMING_SNAKE_CASE =depths _SCREAMING_SNAKE_CASE =hidden_sizes _SCREAMING_SNAKE_CASE =downsampling_rates _SCREAMING_SNAKE_CASE =num_attention_heads _SCREAMING_SNAKE_CASE =is_training _SCREAMING_SNAKE_CASE =use_labels _SCREAMING_SNAKE_CASE =hidden_act _SCREAMING_SNAKE_CASE =hidden_dropout_prob _SCREAMING_SNAKE_CASE =attention_probs_dropout_prob _SCREAMING_SNAKE_CASE =initializer_range _SCREAMING_SNAKE_CASE =num_labels _SCREAMING_SNAKE_CASE =scope def __UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE =None if self.use_labels: _SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _SCREAMING_SNAKE_CASE =self.get_config() return config, pixel_values, labels def __UpperCamelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def __UpperCamelCase ( self : Dict , _a : Dict , _a : Optional[int] , _a : int ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =SegformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def __UpperCamelCase ( self : str , _a : Dict , _a : Optional[int] , _a : Optional[Any] ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =self.num_labels _SCREAMING_SNAKE_CASE =SegformerForSemanticSegmentation(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def __UpperCamelCase ( self : int , _a : int , _a : Optional[Any] , _a : str ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =1 _SCREAMING_SNAKE_CASE =SegformerForSemanticSegmentation(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _SCREAMING_SNAKE_CASE =torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertGreater(result.loss , 0.0 ) def __UpperCamelCase ( self : str ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE =config_and_inputs _SCREAMING_SNAKE_CASE ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): UpperCAmelCase = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) UpperCAmelCase = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False def __UpperCamelCase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =SegformerModelTester(self ) _SCREAMING_SNAKE_CASE =SegformerConfigTester(self , config_class=UpperCamelCase__ ) def __UpperCamelCase ( self : str ) -> int: """simple docstring""" self.config_tester.run_common_tests() def __UpperCamelCase ( self : Dict ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def __UpperCamelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*UpperCamelCase__ ) def __UpperCamelCase ( self : Union[str, Any] ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*UpperCamelCase__ ) @unittest.skip('''SegFormer does not use inputs_embeds''' ) def __UpperCamelCase ( self : Dict ) -> int: """simple docstring""" pass @unittest.skip('''SegFormer does not have get_input_embeddings method and get_output_embeddings methods''' ) def __UpperCamelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" pass def __UpperCamelCase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _SCREAMING_SNAKE_CASE =[*signature.parameters.keys()] _SCREAMING_SNAKE_CASE =["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def __UpperCamelCase ( self : str ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE =True _SCREAMING_SNAKE_CASE =False _SCREAMING_SNAKE_CASE =True _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) _SCREAMING_SNAKE_CASE =outputs.attentions _SCREAMING_SNAKE_CASE =sum(self.model_tester.depths ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _SCREAMING_SNAKE_CASE =True _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) _SCREAMING_SNAKE_CASE =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // 4) ** 2 _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // 32) ** 2 _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) _SCREAMING_SNAKE_CASE =len(UpperCamelCase__ ) # Check attention is always last and order is fine _SCREAMING_SNAKE_CASE =True _SCREAMING_SNAKE_CASE =True _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(out_len + 1 , len(UpperCamelCase__ ) ) _SCREAMING_SNAKE_CASE =outputs.attentions self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first attentions (first block, first layer) _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // 4) ** 2 _SCREAMING_SNAKE_CASE =(self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def __UpperCamelCase ( self : List[Any] ) -> str: """simple docstring""" def check_hidden_states_output(_a : Optional[Any] , _a : Any , _a : Tuple ): _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) _SCREAMING_SNAKE_CASE =outputs.hidden_states _SCREAMING_SNAKE_CASE =self.model_tester.num_encoder_blocks self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _SCREAMING_SNAKE_CASE =True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def __UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" if not self.model_tester.is_training: return _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: if model_class in get_values(UpperCamelCase__ ): continue _SCREAMING_SNAKE_CASE =model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.train() _SCREAMING_SNAKE_CASE =self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =model(**UpperCamelCase__ ).loss loss.backward() @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __UpperCamelCase ( self : Tuple ) -> List[str]: """simple docstring""" pass @slow def __UpperCamelCase ( self : Tuple ) -> str: """simple docstring""" for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =SegformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowerCamelCase( ): _SCREAMING_SNAKE_CASE =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') return image @require_torch class A__ ( unittest.TestCase ): @slow def __UpperCamelCase ( self : Tuple ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =SegformerForSemanticSegmentation.from_pretrained('''nvidia/segformer-b0-finetuned-ade-512-512''' ).to( UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =prepare_img() _SCREAMING_SNAKE_CASE =image_processor(images=UpperCamelCase__ , return_tensors='''pt''' ) _SCREAMING_SNAKE_CASE =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-4 ) ) @slow def __UpperCamelCase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =SegformerForSemanticSegmentation.from_pretrained( '''nvidia/segformer-b1-finetuned-cityscapes-1024-1024''' ).to(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =prepare_img() _SCREAMING_SNAKE_CASE =image_processor(images=UpperCamelCase__ , return_tensors='''pt''' ) _SCREAMING_SNAKE_CASE =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , UpperCamelCase__ , atol=1E-1 ) ) @slow def __UpperCamelCase ( self : int ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase__ , align=UpperCamelCase__ , do_random_crop=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =SegformerForSemanticSegmentation.from_pretrained('''nvidia/segformer-b0-finetuned-ade-512-512''' ).to( UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =prepare_img() _SCREAMING_SNAKE_CASE =image_processor(images=UpperCamelCase__ , return_tensors='''pt''' ) _SCREAMING_SNAKE_CASE =encoded_inputs.pixel_values.to(UpperCamelCase__ ) with torch.no_grad(): _SCREAMING_SNAKE_CASE =model(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =outputs.logits.detach().cpu() _SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ , target_sizes=[(500, 300)] ) _SCREAMING_SNAKE_CASE =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE =torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , UpperCamelCase__ )
691
"""simple docstring""" import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig __A : Optional[Any] = logging.get_logger(__name__) # General docstring __A : str = "PoolFormerConfig" # Base docstring __A : Optional[Any] = "sail/poolformer_s12" __A : List[Any] = [1, 512, 7, 7] # Image classification docstring __A : List[str] = "sail/poolformer_s12" __A : Tuple = "tabby, tabby cat" __A : Tuple = [ "sail/poolformer_s12", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def lowercase ( UpperCamelCase : Any , UpperCamelCase : float = 0.0 , UpperCamelCase : bool = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input A__ : Tuple =1 - drop_prob A__ : List[str] =(input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets A__ : Any =keep_prob + torch.rand(UpperCamelCase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize A__ : Optional[int] =input.div(UpperCamelCase ) * random_tensor return output class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : Optional[float] = None ): super().__init__() A__ : Optional[int] =drop_prob def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : torch.Tensor ): return drop_path(UpperCamelCase__ , self.drop_prob , self.training ) def _UpperCAmelCase ( self : List[str] ): return "p={}".format(self.drop_prob ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=None ): super().__init__() A__ : Optional[int] =patch_size if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (patch_size, patch_size) A__ : Optional[int] =stride if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (stride, stride) A__ : int =padding if isinstance(UpperCamelCase__ , collections.abc.Iterable ) else (padding, padding) A__ : Any =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=UpperCamelCase__ ) A__ : Any =norm_layer(UpperCamelCase__ ) if norm_layer else nn.Identity() def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : str ): A__ : List[str] =self.projection(UpperCamelCase__ ) A__ : Any =self.norm(UpperCamelCase__ ) return embeddings class __lowerCAmelCase ( nn.GroupNorm): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Dict , **UpperCamelCase__ : Union[str, Any] ): super().__init__(1 , UpperCamelCase__ , **UpperCamelCase__ ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Optional[int] ): super().__init__() A__ : Any =nn.AvgPoolad(UpperCamelCase__ , stride=1 , padding=pool_size // 2 , count_include_pad=UpperCamelCase__ ) def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : List[str] ): return self.pool(UpperCamelCase__ ) - hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] ): super().__init__() A__ : List[Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Union[str, Any] =nn.Convad(UpperCamelCase__ , UpperCamelCase__ , 1 ) A__ : Dict =PoolFormerDropPath(UpperCamelCase__ ) if isinstance(config.hidden_act , UpperCamelCase__ ): A__ : Tuple =ACTaFN[config.hidden_act] else: A__ : Optional[Any] =config.hidden_act def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict ): A__ : Optional[Any] =self.conva(UpperCamelCase__ ) A__ : List[str] =self.act_fn(UpperCamelCase__ ) A__ : List[str] =self.drop(UpperCamelCase__ ) A__ : Optional[int] =self.conva(UpperCamelCase__ ) A__ : Optional[Any] =self.drop(UpperCamelCase__ ) return hidden_states class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Any ): super().__init__() A__ : Optional[int] =PoolFormerPooling(UpperCamelCase__ ) A__ : List[str] =PoolFormerOutput(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) A__ : int =PoolFormerGroupNorm(UpperCamelCase__ ) # Useful for training neural nets A__ : Tuple =PoolFormerDropPath(UpperCamelCase__ ) if drop_path > 0.0 else nn.Identity() A__ : Optional[Any] =config.use_layer_scale if config.use_layer_scale: A__ : List[str] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) A__ : List[Any] =nn.Parameter( config.layer_scale_init_value * torch.ones((UpperCamelCase__) ) , requires_grad=UpperCamelCase__ ) def _UpperCAmelCase ( self : Any , UpperCamelCase__ : Optional[int] ): if self.use_layer_scale: A__ : Optional[int] =self.pooling(self.before_norm(UpperCamelCase__ ) ) A__ : Union[str, Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection A__ : Union[str, Any] =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : Tuple =() A__ : List[str] =self.output(self.after_norm(UpperCamelCase__ ) ) A__ : Optional[Any] =self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection A__ : str =hidden_states + self.drop_path(UpperCamelCase__ ) A__ : List[Any] =(output,) + outputs return outputs else: A__ : Tuple =self.drop_path(self.pooling(self.before_norm(UpperCamelCase__ ) ) ) # First residual connection A__ : Optional[Any] =pooling_output + hidden_states A__ : Tuple =() # Second residual connection inside the PoolFormerOutput block A__ : List[str] =self.drop_path(self.output(self.after_norm(UpperCamelCase__ ) ) ) A__ : Any =hidden_states + layer_output A__ : Tuple =(output,) + outputs return outputs class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : List[str] ): super().__init__() A__ : Tuple =config # stochastic depth decay rule A__ : Dict =[x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings A__ : Tuple =[] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) A__ : List[str] =nn.ModuleList(UpperCamelCase__ ) # Transformer blocks A__ : Union[str, Any] =[] A__ : Any =0 for i in range(config.num_encoder_blocks ): # each block consists of layers A__ : Union[str, Any] =[] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( UpperCamelCase__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(UpperCamelCase__ ) ) A__ : str =nn.ModuleList(UpperCamelCase__ ) def _UpperCAmelCase ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : Optional[int]=True ): A__ : Union[str, Any] =() if output_hidden_states else None A__ : Dict =pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): A__ , A__ : List[Any] =layers # Get patch embeddings from hidden_states A__ : Any =embedding_layer(UpperCamelCase__ ) # Send the embeddings through the blocks for _, blk in enumerate(UpperCamelCase__ ): A__ : List[str] =blk(UpperCamelCase__ ) A__ : Tuple =layer_outputs[0] if output_hidden_states: A__ : List[Any] =all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' __magic_name__ : List[str] = PoolFormerConfig __magic_name__ : int = """poolformer""" __magic_name__ : Any = """pixel_values""" __magic_name__ : Any = True def _UpperCAmelCase ( self : List[str] , UpperCamelCase__ : str ): if isinstance(UpperCamelCase__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def _UpperCAmelCase ( self : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any]=False ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A__ : Optional[Any] =value __A : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n" @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : Dict ): super().__init__(UpperCamelCase__ ) A__ : List[Any] =config A__ : Optional[Any] =PoolFormerEncoder(UpperCamelCase__ ) # Initialize weights and apply final processing self.post_init() def _UpperCAmelCase ( self : Tuple ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _UpperCAmelCase ( self : str , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : int =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A__ : Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) A__ : List[Any] =self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : int =encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) class __lowerCAmelCase ( nn.Module): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[Any] ): super().__init__() A__ : List[str] =nn.Linear(config.hidden_size , config.hidden_size ) def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] ): A__ : int =self.dense(UpperCamelCase__ ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , _UpperCamelCase , ) class __lowerCAmelCase ( _UpperCamelCase): '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : str ): super().__init__(UpperCamelCase__ ) A__ : List[str] =config.num_labels A__ : Optional[int] =PoolFormerModel(UpperCamelCase__ ) # Final norm A__ : Dict =PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head A__ : Dict =( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _UpperCAmelCase ( self : Optional[int] , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.LongTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ): A__ : Tuple =return_dict if return_dict is not None else self.config.use_return_dict A__ : List[str] =self.poolformer( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ , ) A__ : str =outputs[0] A__ : List[Any] =self.classifier(self.norm(UpperCamelCase__ ).mean([-2, -1] ) ) A__ : Optional[Any] =None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: A__ : int ="regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): A__ : Tuple ="single_label_classification" else: A__ : Optional[int] ="multi_label_classification" if self.config.problem_type == "regression": A__ : Dict =MSELoss() if self.num_labels == 1: A__ : Optional[Any] =loss_fct(logits.squeeze() , labels.squeeze() ) else: A__ : List[str] =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": A__ : Tuple =CrossEntropyLoss() A__ : int =loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": A__ : List[Any] =BCEWithLogitsLoss() A__ : str =loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: A__ : Optional[int] =(logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states )
656
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCamelCase : Any = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): '''simple docstring''' _UpperCAmelCase : Tuple = """megatron-bert""" def __init__( self : Tuple , lowercase : Dict=29_056 , lowercase : int=1_024 , lowercase : Optional[int]=24 , lowercase : Dict=16 , lowercase : int=4_096 , lowercase : str="gelu" , lowercase : List[str]=0.1 , lowercase : int=0.1 , lowercase : int=512 , lowercase : str=2 , lowercase : Union[str, Any]=0.02 , lowercase : Any=1E-12 , lowercase : List[Any]=0 , lowercase : str="absolute" , lowercase : Dict=True , **lowercase : Tuple , ): '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = position_embedding_type _snake_case = use_cache
686
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase): '''simple docstring''' __magic_name__ : int = IFInpaintingSuperResolutionPipeline __magic_name__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} __magic_name__ : str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""}) __magic_name__ : Optional[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def _UpperCAmelCase ( self : Union[str, Any] ): return self._get_superresolution_dummy_components() def _UpperCAmelCase ( self : Optional[Any] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[int]=0 ): if str(UpperCamelCase__ ).startswith("mps" ): A__ : Any =torch.manual_seed(UpperCamelCase__ ) else: A__ : Dict =torch.Generator(device=UpperCamelCase__ ).manual_seed(UpperCamelCase__ ) A__ : Tuple =floats_tensor((1, 3, 16, 16) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Optional[int] =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : Any =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCamelCase__ ) ).to(UpperCamelCase__ ) A__ : List[str] ={ "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _UpperCAmelCase ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _UpperCAmelCase ( self : int ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _UpperCAmelCase ( self : Tuple ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def _UpperCAmelCase ( self : str ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _UpperCAmelCase ( self : Dict ): self._test_save_load_local() def _UpperCAmelCase ( self : Optional[int] ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
656
0
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset from utils import logger class _lowerCAmelCase ( lowerCamelCase ): def __init__( self , a_ , a_ ) -> Optional[Any]: _UpperCAmelCase = params _UpperCAmelCase = np.array(a_ ) _UpperCAmelCase = np.array([len(a_ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self , a_ ) -> Any: return (self.token_ids[index], self.lengths[index]) def __len__( self ) -> List[Any]: return len(self.lengths ) def _a ( self ) -> Optional[int]: assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = self.params.max_model_input_size _UpperCAmelCase = self.lengths > max_len logger.info(f"Splitting {sum(a_ )} too long sequences." ) def divide_chunks(a_ , a_ ): return [l[i : i + n] for i in range(0 , len(a_ ) , a_ )] _UpperCAmelCase = [] _UpperCAmelCase = [] if self.params.mlm: _UpperCAmelCase , _UpperCAmelCase = self.params.special_tok_ids["cls_token"], self.params.special_tok_ids["sep_token"] else: _UpperCAmelCase , _UpperCAmelCase = self.params.special_tok_ids["bos_token"], self.params.special_tok_ids["eos_token"] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: _UpperCAmelCase = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: _UpperCAmelCase = np.insert(a_ , 0 , a_ ) if sub_s[-1] != sep_id: _UpperCAmelCase = np.insert(a_ , len(a_ ) , a_ ) assert len(a_ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a_ ) new_tok_ids.extend(a_ ) new_lengths.extend([len(a_ ) for l in sub_seqs] ) _UpperCAmelCase = np.array(a_ ) _UpperCAmelCase = np.array(a_ ) def _a ( self ) -> List[Any]: _UpperCAmelCase = len(self ) _UpperCAmelCase = self.lengths > 11 _UpperCAmelCase = self.token_ids[indices] _UpperCAmelCase = self.lengths[indices] _UpperCAmelCase = len(self ) logger.info(f"Remove {init_size - new_size} too short (<=11 tokens) sequences." ) def _a ( self ) -> Dict: if "unk_token" not in self.params.special_tok_ids: return else: _UpperCAmelCase = self.params.special_tok_ids["unk_token"] _UpperCAmelCase = len(self ) _UpperCAmelCase = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) _UpperCAmelCase = (unk_occs / self.lengths) < 0.5 _UpperCAmelCase = self.token_ids[indices] _UpperCAmelCase = self.lengths[indices] _UpperCAmelCase = len(self ) logger.info(f"Remove {init_size - new_size} sequences with a high level of unknown tokens (50%)." ) def _a ( self ) -> Optional[int]: if not self.params.is_master: return logger.info(f"{len(self )} sequences" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _a ( self , a_ ) -> Dict: _UpperCAmelCase = [t[0] for t in batch] _UpperCAmelCase = [t[1] for t in batch] assert len(a_ ) == len(a_ ) # Max for paddings _UpperCAmelCase = max(a_ ) # Pad token ids if self.params.mlm: _UpperCAmelCase = self.params.special_tok_ids["pad_token"] else: _UpperCAmelCase = self.params.special_tok_ids["unk_token"] _UpperCAmelCase = [list(t.astype(a_ ) ) + [pad_idx] * (max_seq_len_ - len(a_ )) for t in token_ids] assert len(tk_ ) == len(a_ ) assert all(len(a_ ) == max_seq_len_ for t in tk_ ) _UpperCAmelCase = torch.tensor(tk_ ) # (bs, max_seq_len_) _UpperCAmelCase = torch.tensor(a_ ) # (bs) return tk_t, lg_t
657
"""simple docstring""" import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _lowerCAmelCase : def __init__( self , a_ , a_=2 , a_=3 , a_=4 , a_=2 , a_=7 , a_=True , a_=True , a_=True , a_=True , a_=99 , a_=36 , a_=3 , a_=4 , a_=37 , a_="gelu" , a_=0.1 , a_=0.1 , a_=512 , a_=16 , a_=2 , a_=0.02 , a_=6 , a_=6 , a_=3 , a_=4 , a_=None , a_=1000 , ) -> Optional[Any]: _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = num_channels _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = text_seq_length _UpperCAmelCase = is_training _UpperCAmelCase = use_input_mask _UpperCAmelCase = use_token_type_ids _UpperCAmelCase = use_labels _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = type_vocab_size _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = initializer_range _UpperCAmelCase = coordinate_size _UpperCAmelCase = shape_size _UpperCAmelCase = num_labels _UpperCAmelCase = num_choices _UpperCAmelCase = scope _UpperCAmelCase = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _UpperCAmelCase = text_seq_length _UpperCAmelCase = (image_size // patch_size) ** 2 + 1 _UpperCAmelCase = self.text_seq_length + self.image_seq_length def _a ( self ) -> Dict: _UpperCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _UpperCAmelCase = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _UpperCAmelCase = bbox[i, j, 3] _UpperCAmelCase = bbox[i, j, 1] _UpperCAmelCase = t if bbox[i, j, 2] < bbox[i, j, 0]: _UpperCAmelCase = bbox[i, j, 2] _UpperCAmelCase = bbox[i, j, 0] _UpperCAmelCase = t _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_input_mask: _UpperCAmelCase = random_attention_mask([self.batch_size, self.text_seq_length] ) _UpperCAmelCase = None if self.use_token_type_ids: _UpperCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _UpperCAmelCase = None _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _UpperCAmelCase = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def _a ( self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ ) -> Tuple: _UpperCAmelCase = LayoutLMvaModel(config=a_ ) model.to(a_ ) model.eval() # text + image _UpperCAmelCase = model(a_ , pixel_values=a_ ) _UpperCAmelCase = model( a_ , bbox=a_ , pixel_values=a_ , attention_mask=a_ , token_type_ids=a_ ) _UpperCAmelCase = model(a_ , bbox=a_ , pixel_values=a_ , token_type_ids=a_ ) _UpperCAmelCase = model(a_ , bbox=a_ , pixel_values=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _UpperCAmelCase = model(a_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _UpperCAmelCase = model(pixel_values=a_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def _a ( self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[Any]: _UpperCAmelCase = self.num_labels _UpperCAmelCase = LayoutLMvaForSequenceClassification(a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model( a_ , bbox=a_ , pixel_values=a_ , attention_mask=a_ , token_type_ids=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ ) -> Union[str, Any]: _UpperCAmelCase = self.num_labels _UpperCAmelCase = LayoutLMvaForTokenClassification(config=a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model( a_ , bbox=a_ , pixel_values=a_ , attention_mask=a_ , token_type_ids=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def _a ( self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ ) -> Dict: _UpperCAmelCase = LayoutLMvaForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model( a_ , bbox=a_ , pixel_values=a_ , attention_mask=a_ , token_type_ids=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a ( self ) -> Optional[int]: _UpperCAmelCase = self.prepare_config_and_inputs() ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) = config_and_inputs _UpperCAmelCase = { "input_ids": input_ids, "bbox": bbox, "pixel_values": pixel_values, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_torch class _lowerCAmelCase ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase_ : Any = False lowercase_ : Dict = False lowercase_ : List[str] = False lowercase_ : str = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) lowercase_ : int = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def _a ( self , a_ , a_ , a_ , a_ , a_ ) -> List[str]: # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = LayoutLMvaModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=a_ , hidden_size=37 ) def _a ( self , a_ , a_ , a_=False ) -> List[str]: _UpperCAmelCase = copy.deepcopy(a_ ) if model_class in get_values(a_ ): _UpperCAmelCase = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(a_ , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(a_ ): _UpperCAmelCase = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=a_ ) elif model_class in get_values(a_ ): _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a_ ) _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a_ ) elif model_class in [ *get_values(a_ ), ]: _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a_ ) elif model_class in [ *get_values(a_ ), ]: _UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=a_ , ) return inputs_dict def _a ( self ) -> int: self.config_tester.run_common_tests() def _a ( self ) -> List[str]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def _a ( self ) -> List[str]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _UpperCAmelCase = type self.model_tester.create_and_check_model(*a_ ) def _a ( self ) -> int: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a_ ) def _a ( self ) -> Any: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a_ ) def _a ( self ) -> List[Any]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a_ ) @slow def _a ( self ) -> List[str]: for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = LayoutLMvaModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class _lowerCAmelCase ( unittest.TestCase ): @cached_property def _a ( self ) -> List[Any]: return LayoutLMvaImageProcessor(apply_ocr=a_ ) if is_vision_available() else None @slow def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = LayoutLMvaModel.from_pretrained("microsoft/layoutlmv3-base" ).to(a_ ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=a_ , return_tensors="pt" ).pixel_values.to(a_ ) _UpperCAmelCase = torch.tensor([[1, 2]] ) _UpperCAmelCase = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _UpperCAmelCase = model( input_ids=input_ids.to(a_ ) , bbox=bbox.to(a_ ) , pixel_values=pixel_values.to(a_ ) , ) # verify the logits _UpperCAmelCase = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , a_ ) _UpperCAmelCase = torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ).to(a_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , a_ , atol=1e-4 ) )
657
1
"""simple docstring""" import math def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" return math.pow(UpperCamelCase__ , 2 ) - a def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" return 2 * x def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = 2.0 while start <= a: _UpperCAmelCase = math.pow(UpperCamelCase__ , 2 ) return start def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ = 9999 , UpperCamelCase__ = 0.00000000000001 ): """simple docstring""" if a < 0: raise ValueError("math domain error" ) _UpperCAmelCase = get_initial_point(UpperCamelCase__ ) for _ in range(UpperCamelCase__ ): _UpperCAmelCase = value _UpperCAmelCase = value - fx(UpperCamelCase__ , UpperCamelCase__ ) / fx_derivative(UpperCamelCase__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
657
"""simple docstring""" import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _lowerCAmelCase ( unittest.TestCase ): lowercase_ : str = MODEL_FOR_MASKED_LM_MAPPING lowercase_ : List[str] = TF_MODEL_FOR_MASKED_LM_MAPPING def _a ( self ) -> Optional[Any]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def _a ( self ) -> str: _UpperCAmelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf" ) _UpperCAmelCase = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ {"sequence": "My name is grouped", "score": 2.1e-05, "token": 38015, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1e-05, "token": 25506, "token_str": " accuser"}, ] , ) _UpperCAmelCase = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ { "sequence": "The largest city in France is grouped", "score": 2.1e-05, "token": 38015, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1e-05, "token": 25506, "token_str": " accuser", }, ] , ) _UpperCAmelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ {"sequence": "My name is Clara", "score": 2e-05, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2e-05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9e-05, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt" ) _UpperCAmelCase = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ {"sequence": "My name is Maul", "score": 2.2e-05, "token": 35676, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2e-05, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCAmelCase = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ { "sequence": "The largest city in France is Maul", "score": 2.2e-05, "token": 35676, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2e-05, "token": 16416, "token_str": "ELS"}, ] , ) _UpperCAmelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ {"sequence": "My name is Patrick", "score": 2.1e-05, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2e-05, "token": 2941, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2e-05, "token": 13606, "token_str": " Clara"}, ] , ) _UpperCAmelCase = unmasker("My name is <mask> <mask>" , top_k=2 ) self.assertEqual( nested_simplify(a_ , decimals=6 ) , [ [ { "score": 2.2e-05, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2e-05, "token": 16416, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2e-05, "token": 35676, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2e-05, "token": 16416, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def _a ( self ) -> int: _UpperCAmelCase = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt" ) # convert model to fp16 pipe.model.half() _UpperCAmelCase = pipe("Paris is the [MASK] of France." ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(a_ , a_ ) @slow @require_torch def _a ( self ) -> int: _UpperCAmelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt" ) self.run_large_test(a_ ) @slow @require_tf def _a ( self ) -> int: _UpperCAmelCase = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf" ) self.run_large_test(a_ ) def _a ( self , a_ ) -> int: _UpperCAmelCase = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(a_ ) , [ {"sequence": "My name is John", "score": 0.008, "token": 610, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.007, "token": 1573, "token_str": " Chris"}, ] , ) _UpperCAmelCase = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(a_ ) , [ { "sequence": "The largest city in France is Paris", "score": 0.251, "token": 2201, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.214, "token": 12790, "token_str": " Lyon", }, ] , ) _UpperCAmelCase = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(a_ ) , [ {"sequence": "My name is Patrick", "score": 0.005, "token": 3499, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.000, "token": 13606, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.000, "token": 2941, "token_str": " Te"}, ] , ) @require_torch def _a ( self ) -> Any: _UpperCAmelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt" ) _UpperCAmelCase = None _UpperCAmelCase = None self.run_pipeline_test(a_ , [] ) @require_tf def _a ( self ) -> List[Any]: _UpperCAmelCase = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf" ) _UpperCAmelCase = None _UpperCAmelCase = None self.run_pipeline_test(a_ , [] ) def _a ( self , a_ , a_ , a_ ) -> Optional[Any]: if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)" ) _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) _UpperCAmelCase = [ f"This is another {tokenizer.mask_token} test", ] return fill_masker, examples def _a ( self , a_ , a_ ) -> List[str]: _UpperCAmelCase = fill_masker.tokenizer _UpperCAmelCase = fill_masker.model _UpperCAmelCase = fill_masker( f"This is a {tokenizer.mask_token}" , ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) _UpperCAmelCase = fill_masker([f"This is a {tokenizer.mask_token}"] ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) _UpperCAmelCase = fill_masker([f"This is a {tokenizer.mask_token}", f"Another {tokenizer.mask_token} great test."] ) self.assertEqual( a_ , [ [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ], [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ], ] , ) with self.assertRaises(a_ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(a_ ): fill_masker("This is" ) self.run_test_top_k(a_ , a_ ) self.run_test_targets(a_ , a_ ) self.run_test_top_k_targets(a_ , a_ ) self.fill_mask_with_duplicate_targets_and_top_k(a_ , a_ ) self.fill_mask_with_multiple_masks(a_ , a_ ) def _a ( self , a_ , a_ ) -> Optional[int]: _UpperCAmelCase = tokenizer.get_vocab() _UpperCAmelCase = sorted(vocab.keys() )[:2] # Pipeline argument _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ , targets=a_ ) _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) _UpperCAmelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , a_ ) _UpperCAmelCase = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(a_ ) ) # Call argument _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets=a_ ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) _UpperCAmelCase = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , a_ ) _UpperCAmelCase = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(a_ ) ) # Score equivalence _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets=a_ ) _UpperCAmelCase = [top_mask["token_str"] for top_mask in outputs] _UpperCAmelCase = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(a_ ) == set(a_ ): _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets=a_ ) _UpperCAmelCase = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(a_ ) , nested_simplify(a_ ) ) # Raises with invalid with self.assertRaises(a_ ): _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(a_ ): _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets=[""] ) with self.assertRaises(a_ ): _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , targets="" ) def _a ( self , a_ , a_ ) -> str: _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ , top_k=2 ) _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( a_ , [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ] , ) self.assertEqual(nested_simplify(a_ ) , nested_simplify(a_ ) ) def _a ( self , a_ , a_ ) -> List[Any]: _UpperCAmelCase = tokenizer.get_vocab() _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) # top_k=2, ntargets=3 _UpperCAmelCase = sorted(vocab.keys() )[:3] _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , top_k=2 , targets=a_ ) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCAmelCase = [el["token_str"] for el in sorted(a_ , key=lambda a_ : x["score"] , reverse=a_ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(a_ ).issubset(a_ ): _UpperCAmelCase = fill_masker(f"This is a {tokenizer.mask_token}" , top_k=3 , targets=a_ ) # They should yield exactly the same result self.assertEqual(nested_simplify(a_ ) , nested_simplify(a_ ) ) def _a ( self , a_ , a_ ) -> Optional[Any]: _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) _UpperCAmelCase = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCAmelCase = sorted(vocab.keys() )[:3] _UpperCAmelCase = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCAmelCase = fill_masker(f"My name is {tokenizer.mask_token}" , targets=a_ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(a_ ) , 3 ) def _a ( self , a_ , a_ ) -> Any: _UpperCAmelCase = FillMaskPipeline(model=a_ , tokenizer=a_ ) _UpperCAmelCase = fill_masker( f"This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( a_ , [ [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ], [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ], [ {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, {"sequence": ANY(a_ ), "score": ANY(a_ ), "token": ANY(a_ ), "token_str": ANY(a_ )}, ], ] , )
657
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class _lowerCAmelCase ( unittest.TestCase ): @require_torch def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = pipeline( task="zero-shot-audio-classification" , model="hf-internal-testing/tiny-clap-htsat-unfused" ) _UpperCAmelCase = load_dataset("ashraq/esc50" ) _UpperCAmelCase = dataset["train"]["audio"][-1]["array"] _UpperCAmelCase = audio_classifier(a_ , candidate_labels=["Sound of a dog", "Sound of vaccum cleaner"] ) self.assertEqual( nested_simplify(a_ ) , [{"score": 0.501, "label": "Sound of a dog"}, {"score": 0.499, "label": "Sound of vaccum cleaner"}] , ) @unittest.skip("No models are available in TF" ) def _a ( self ) -> Tuple: pass @slow @require_torch def _a ( self ) -> List[Any]: _UpperCAmelCase = pipeline( task="zero-shot-audio-classification" , model="laion/clap-htsat-unfused" , ) # This is an audio of a dog _UpperCAmelCase = load_dataset("ashraq/esc50" ) _UpperCAmelCase = dataset["train"]["audio"][-1]["array"] _UpperCAmelCase = audio_classifier(a_ , candidate_labels=["Sound of a dog", "Sound of vaccum cleaner"] ) self.assertEqual( nested_simplify(a_ ) , [ {"score": 0.999, "label": "Sound of a dog"}, {"score": 0.001, "label": "Sound of vaccum cleaner"}, ] , ) _UpperCAmelCase = audio_classifier([audio] * 5 , candidate_labels=["Sound of a dog", "Sound of vaccum cleaner"] ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.999, "label": "Sound of a dog"}, {"score": 0.001, "label": "Sound of vaccum cleaner"}, ], ] * 5 , ) _UpperCAmelCase = audio_classifier( [audio] * 5 , candidate_labels=["Sound of a dog", "Sound of vaccum cleaner"] , batch_size=5 ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.999, "label": "Sound of a dog"}, {"score": 0.001, "label": "Sound of vaccum cleaner"}, ], ] * 5 , ) @unittest.skip("No models are available in TF" ) def _a ( self ) -> Dict: pass
657
"""simple docstring""" import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class _lowerCAmelCase ( lowerCamelCase ): def _a ( self ) -> List[str]: _UpperCAmelCase = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def _a ( self ) -> Optional[int]: with self.assertRaises(a_ ): _UpperCAmelCase = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def _a ( self ) -> int: with self.assertRaises(a_ ): _UpperCAmelCase = pa.array(TypedSequence([1, 2, 3] , try_type=Value("bool" ) , type=Value("int64" ) ) ) def _a ( self ) -> Optional[Any]: _UpperCAmelCase = pa.array(TypedSequence([1, 2, 3] , type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _a ( self ) -> int: with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): _UpperCAmelCase = pa.array(TypedSequence(["foo", "bar"] , type=Value("int64" ) ) ) def _a ( self ) -> Dict: _UpperCAmelCase = pa.array(TypedSequence([1, 2, 3] , try_type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = pa.array(TypedSequence(["foo", "bar"] , try_type=Value("int64" ) ) ) self.assertEqual(arr.type , pa.string() ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def _a ( self ) -> Tuple: with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): _UpperCAmelCase = pa.array(TypedSequence(["foo", "bar"] , type=ArrayaD((1, 3) , "int64" ) ) ) def _a ( self ) -> str: _UpperCAmelCase = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def _a ( self ) -> Tuple: _UpperCAmelCase = pa.array(TypedSequence(["foo", "bar"] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def _a ( self ) -> List[str]: import PIL.Image _UpperCAmelCase = PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( "datasets.arrow_writer.cast_to_python_objects" , side_effect=a_ ) as mock_cast_to_python_objects: _UpperCAmelCase = pa.array(TypedSequence([{"path": None, "bytes": B"image_bytes"}, pil_image] , type=Image() ) ) _UpperCAmelCase , _UpperCAmelCase = mock_cast_to_python_objects.call_args_list[-1] self.assertIn("optimize_list_casting" , a_ ) self.assertFalse(kwargs["optimize_list_casting"] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferReader(UpperCamelCase__ ) if isinstance(UpperCamelCase__ , pa.Buffer ) else pa.memory_map(UpperCamelCase__ ) _UpperCAmelCase = pa.ipc.open_stream(UpperCamelCase__ ) _UpperCAmelCase = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() _UpperCAmelCase = pa.schema(UpperCamelCase__ ) if fields else None with ArrowWriter(stream=UpperCamelCase__ , schema=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: _UpperCAmelCase = {"col_1": pa.string(), "col_2": pa.intaa()} assert writer._schema == pa.schema(UpperCamelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() _UpperCAmelCase = Features({"labels": ClassLabel(names=["neg", "pos"] )} ) with ArrowWriter(stream=UpperCamelCase__ , features=UpperCamelCase__ ) as writer: writer.write({"labels": 0} ) writer.write({"labels": 1} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata _UpperCAmelCase = pa.BufferReader(output.getvalue() ) _UpperCAmelCase = pa.ipc.open_stream(UpperCamelCase__ ) _UpperCAmelCase = f.read_all() _UpperCAmelCase = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(UpperCamelCase__ ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ , hash_salt="split_name" , check_duplicates=UpperCamelCase__ , ) as writer: with pytest.raises(UpperCamelCase__ ): writer.write({"col_1": "foo", "col_2": 1} , key=[1, 2] ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 10] ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ , hash_salt="split_name" , check_duplicates=UpperCamelCase__ , ) as writer: with pytest.raises(UpperCamelCase__ ): writer.write({"col_1": "foo", "col_2": 1} , key=10 ) writer.write({"col_1": "bar", "col_2": 2} , key=10 ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 10] ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() with ArrowWriter( stream=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ , hash_salt="split_name" , check_duplicates=UpperCamelCase__ , ) as writer: writer.write({"col_1": "foo", "col_2": 1} , key=1 ) writer.write({"col_1": "bar", "col_2": 2} , key=2 ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() _UpperCAmelCase = pa.schema(UpperCamelCase__ ) if fields else None with ArrowWriter(stream=UpperCamelCase__ , schema=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) writer.write_batch({"col_1": [], "col_2": []} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: _UpperCAmelCase = {"col_1": pa.string(), "col_2": pa.intaa()} assert writer._schema == pa.schema(UpperCamelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() _UpperCAmelCase = pa.schema(UpperCamelCase__ ) if fields else None with ArrowWriter(stream=UpperCamelCase__ , schema=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ ) as writer: writer.write_table(pa.Table.from_pydict({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: _UpperCAmelCase = {"col_1": pa.string(), "col_2": pa.intaa()} assert writer._schema == pa.schema(UpperCamelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 10] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() _UpperCAmelCase = pa.schema(UpperCamelCase__ ) if fields else None with ArrowWriter(stream=UpperCamelCase__ , schema=UpperCamelCase__ , writer_batch_size=UpperCamelCase__ ) as writer: writer.write_row(pa.Table.from_pydict({"col_1": ["foo"], "col_2": [1]} ) ) writer.write_row(pa.Table.from_pydict({"col_1": ["bar"], "col_2": [2]} ) ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: _UpperCAmelCase = {"col_1": pa.string(), "col_2": pa.intaa()} assert writer._schema == pa.schema(UpperCamelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def __lowerCamelCase ( ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: _UpperCAmelCase = {"col_1": pa.string(), "col_2": pa.intaa()} _UpperCAmelCase = os.path.join(UpperCamelCase__ , "test.arrow" ) with ArrowWriter(path=UpperCamelCase__ , schema=pa.schema(UpperCamelCase__ ) ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(UpperCamelCase__ , metadata=writer._schema.metadata ) _check_output(UpperCamelCase__ , 1 ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if pa.types.is_list(UpperCamelCase__ ): return get_base_dtype(arr_type.value_type ) else: return arr_type def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if isinstance(lst[0] , UpperCamelCase__ ): change_first_primitive_element_in_list(lst[0] , UpperCamelCase__ ) else: _UpperCAmelCase = value @pytest.mark.parametrize("optimized_int_type, expected_dtype" , [(None, pa.intaa()), (Value("int32" ), pa.intaa())] ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.array(TypedSequence(UpperCamelCase__ , optimized_int_type=UpperCamelCase__ ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( "col, expected_dtype" , [ ("attention_mask", pa.inta()), ("special_tokens_mask", pa.inta()), ("token_type_ids", pa.inta()), ("input_ids", pa.intaa()), ("other", pa.intaa()), ] , ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pa.array(OptimizedTypedSequence(UpperCamelCase__ , col=UpperCamelCase__ ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications _UpperCAmelCase = copy.deepcopy(UpperCamelCase__ ) _UpperCAmelCase = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(UpperCamelCase__ , UpperCamelCase__ ) _UpperCAmelCase = pa.array(OptimizedTypedSequence(UpperCamelCase__ , col=UpperCamelCase__ ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize("raise_exception" , [False, True] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = str(tmp_path / "dataset-train.arrow" ) try: with ArrowWriter(path=UpperCamelCase__ ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = "mock://dataset-train.arrow" with ArrowWriter(path=UpperCamelCase__ , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(UpperCamelCase__ ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(UpperCamelCase__ ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = pa.BufferOutputStream() with ParquetWriter(stream=UpperCamelCase__ ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) _UpperCAmelCase , _UpperCAmelCase = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _UpperCAmelCase = pa.BufferReader(output.getvalue() ) _UpperCAmelCase = pq.read_table(UpperCamelCase__ ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize("embed_local_files" , [False, True] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" import PIL.Image _UpperCAmelCase = str(tmp_path / "test_image_rgb.jpg" ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(UpperCamelCase__ , format="png" ) _UpperCAmelCase = pa.BufferOutputStream() with ParquetWriter( stream=UpperCamelCase__ , features=Features({"image": Image()} ) , embed_local_files=UpperCamelCase__ ) as writer: writer.write({"image": image_path} ) writer.finalize() _UpperCAmelCase = pa.BufferReader(output.getvalue() ) _UpperCAmelCase = pq.read_table(UpperCamelCase__ ) _UpperCAmelCase = pa_table.to_pydict() if embed_local_files: assert isinstance(out["image"][0]["path"] , UpperCamelCase__ ) with open(UpperCamelCase__ , "rb" ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = pa.schema([pa.field("col_1" , pa.string() , nullable=UpperCamelCase__ )] ) _UpperCAmelCase = pa.BufferOutputStream() with ArrowWriter(stream=UpperCamelCase__ ) as writer: writer._build_writer(inferred_schema=UpperCamelCase__ ) assert writer._schema == pa.schema([pa.field("col_1" , pa.string() )] )
657
1
"""simple docstring""" import os from distutils.util import strtobool def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" for e in env_keys: _UpperCAmelCase = int(os.environ.get(UpperCamelCase__ , -1 ) ) if val >= 0: return val return default def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=False ): """simple docstring""" _UpperCAmelCase = os.environ.get(UpperCamelCase__ , str(UpperCamelCase__ ) ) return strtobool(UpperCamelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__="no" ): """simple docstring""" _UpperCAmelCase = os.environ.get(UpperCamelCase__ , str(UpperCamelCase__ ) ) return value
657
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class _lowerCAmelCase ( unittest.TestCase ): def _a ( self ) -> Optional[Any]: _UpperCAmelCase = ["a", "b", "c"] # Defaults to last layer if both are None _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , a_ , a_ ) self.assertEqual(a_ , ["c"] ) self.assertEqual(a_ , [2] ) # Out indices set to match out features _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(["a", "c"] , a_ , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [0, 2] ) # Out features set to match out indices _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , [0, 2] , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [0, 2] ) # Out features selected from negative indices _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , [-3, -1] , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [-3, -1] ) def _a ( self ) -> Optional[int]: # Stage names must be set with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , a_ ) # Out features must be a list with self.assertRaises(a_ ): verify_out_features_out_indices(("a", "b") , (0, 1) , ["a", "b"] ) # Out features must be a subset of stage names with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , ["a"] ) # Out indices must be a list or tuple with self.assertRaises(a_ ): verify_out_features_out_indices(a_ , 0 , ["a", "b"] ) # Out indices must be a subset of stage names with self.assertRaises(a_ ): verify_out_features_out_indices(a_ , (0, 1) , ["a"] ) # Out features and out indices must be the same length with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0,) , ["a", "b", "c"] ) # Out features should match out indices with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 2) , ["a", "b", "c"] ) # Out features and out indices should be in order with self.assertRaises(a_ ): verify_out_features_out_indices(["b", "a"] , (0, 1) , ["a", "b"] ) # Check passes with valid inputs verify_out_features_out_indices(["a", "b", "d"] , (0, 1, -1) , ["a", "b", "c", "d"] ) def _a ( self ) -> int: _UpperCAmelCase = BackboneMixin() _UpperCAmelCase = ["a", "b", "c"] _UpperCAmelCase = ["a", "c"] _UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly _UpperCAmelCase = ["a", "b"] self.assertEqual(backbone.out_features , ["a", "b"] ) self.assertEqual(backbone.out_indices , [0, 1] ) _UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [-3, -1] )
657
1
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib __magic_name__ = get_logger() __magic_name__ = None class _lowerCAmelCase ( TensorFormatter[Mapping, '''jax.Array''', Mapping] ): def __init__( self , a_=None , a_=None , **a_ ) -> int: super().__init__(features=a_ ) import jax from jaxlib.xla_client import Device if isinstance(a_ , a_ ): raise ValueError( f"Expected {device} to be a `str` not {type(a_ )}, as `jaxlib.xla_extension.Device` " "is not serializable neither with `pickle` nor with `dill`. Instead you can surround " "the device with `str()` to get its string identifier that will be internally mapped " "to the actual `jaxlib.xla_extension.Device`." ) _UpperCAmelCase = device if isinstance(a_ , a_ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: _UpperCAmelCase = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( f"Device with string identifier {self.device} not listed among the available " f"devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default " f"device: {str(jax.devices()[0] )}." ) _UpperCAmelCase = str(jax.devices()[0] ) _UpperCAmelCase = jnp_array_kwargs @staticmethod def _a ( ) -> Dict[str, "jaxlib.xla_extension.Device"]: import jax return {str(a_ ): device for device in jax.devices()} def _a ( self , a_ ) -> Optional[int]: import jax import jax.numpy as jnp if isinstance(a_ , a_ ) and column: if all( isinstance(a_ , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(a_ , axis=0 ) return column def _a ( self , a_ ) -> Union[str, Any]: import jax import jax.numpy as jnp if isinstance(a_ , (str, bytes, type(a_ )) ): return value elif isinstance(a_ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() _UpperCAmelCase = {} if isinstance(a_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: _UpperCAmelCase = {"dtype": jnp.intaa} else: _UpperCAmelCase = {"dtype": jnp.intaa} elif isinstance(a_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): _UpperCAmelCase = {"dtype": jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(a_ , PIL.Image.Image ): _UpperCAmelCase = np.asarray(a_ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: _UpperCAmelCase = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(a_ , **{**default_dtype, **self.jnp_array_kwargs} ) def _a ( self , a_ ) -> Union[str, Any]: import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(a_ , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(a_ , "__array__" ) and not isinstance(a_ , jax.Array ): _UpperCAmelCase = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(a_ , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(a_ ) for substruct in data_struct] ) elif isinstance(a_ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(a_ ) for substruct in data_struct] ) return self._tensorize(a_ ) def _a ( self , a_ ) -> List[Any]: return map_nested(self._recursive_tensorize , a_ , map_list=a_ ) def _a ( self , a_ ) -> Mapping: _UpperCAmelCase = self.numpy_arrow_extractor().extract_row(a_ ) _UpperCAmelCase = self.python_features_decoder.decode_row(a_ ) return self.recursive_tensorize(a_ ) def _a ( self , a_ ) -> "jax.Array": _UpperCAmelCase = self.numpy_arrow_extractor().extract_column(a_ ) _UpperCAmelCase = self.python_features_decoder.decode_column(a_ , pa_table.column_names[0] ) _UpperCAmelCase = self.recursive_tensorize(a_ ) _UpperCAmelCase = self._consolidate(a_ ) return column def _a ( self , a_ ) -> Mapping: _UpperCAmelCase = self.numpy_arrow_extractor().extract_batch(a_ ) _UpperCAmelCase = self.python_features_decoder.decode_batch(a_ ) _UpperCAmelCase = self.recursive_tensorize(a_ ) for column_name in batch: _UpperCAmelCase = self._consolidate(batch[column_name] ) return batch
657
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __magic_name__ = { '''configuration_electra''': ['''ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ElectraConfig''', '''ElectraOnnxConfig'''], '''tokenization_electra''': ['''ElectraTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = ['''ElectraTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ElectraForCausalLM''', '''ElectraForMaskedLM''', '''ElectraForMultipleChoice''', '''ElectraForPreTraining''', '''ElectraForQuestionAnswering''', '''ElectraForSequenceClassification''', '''ElectraForTokenClassification''', '''ElectraModel''', '''ElectraPreTrainedModel''', '''load_tf_weights_in_electra''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFElectraForMaskedLM''', '''TFElectraForMultipleChoice''', '''TFElectraForPreTraining''', '''TFElectraForQuestionAnswering''', '''TFElectraForSequenceClassification''', '''TFElectraForTokenClassification''', '''TFElectraModel''', '''TFElectraPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''FlaxElectraForCausalLM''', '''FlaxElectraForMaskedLM''', '''FlaxElectraForMultipleChoice''', '''FlaxElectraForPreTraining''', '''FlaxElectraForQuestionAnswering''', '''FlaxElectraForSequenceClassification''', '''FlaxElectraForTokenClassification''', '''FlaxElectraModel''', '''FlaxElectraPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
657
1
"""simple docstring""" import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( lowerCamelCase , unittest.TestCase ): lowercase_ : List[str] = XLMTokenizer lowercase_ : Optional[int] = False def _a ( self ) -> List[str]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCAmelCase = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] _UpperCAmelCase = dict(zip(a_ , range(len(a_ ) ) ) ) _UpperCAmelCase = ["l o 123", "lo w 1456", "e r</w> 1789", ""] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" ) as fp: fp.write(json.dumps(a_ ) ) with open(self.merges_file , "w" ) as fp: fp.write("\n".join(a_ ) ) def _a ( self , a_ ) -> Tuple: _UpperCAmelCase = "lower newer" _UpperCAmelCase = "lower newer" return input_text, output_text def _a ( self ) -> int: _UpperCAmelCase = XLMTokenizer(self.vocab_file , self.merges_file ) _UpperCAmelCase = "lower" _UpperCAmelCase = ["low", "er</w>"] _UpperCAmelCase = tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) _UpperCAmelCase = tokens + ["<unk>"] _UpperCAmelCase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , a_ ) @slow def _a ( self ) -> str: _UpperCAmelCase = XLMTokenizer.from_pretrained("xlm-mlm-en-2048" ) _UpperCAmelCase = tokenizer.encode("sequence builders" , add_special_tokens=a_ ) _UpperCAmelCase = tokenizer.encode("multi-sequence build" , add_special_tokens=a_ ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(a_ ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(a_ , a_ ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
657
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class _lowerCAmelCase ( lowerCamelCase , unittest.TestCase ): lowercase_ : Tuple = BarthezTokenizer lowercase_ : List[Any] = BarthezTokenizerFast lowercase_ : Dict = True lowercase_ : int = True def _a ( self ) -> Any: super().setUp() _UpperCAmelCase = BarthezTokenizerFast.from_pretrained("moussaKam/mbarthez" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=a_ ) _UpperCAmelCase = tokenizer def _a ( self ) -> List[Any]: _UpperCAmelCase = "<pad>" _UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def _a ( self ) -> List[Any]: _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(a_ ) , 101122 ) def _a ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 101122 ) @require_torch def _a ( self ) -> List[Any]: _UpperCAmelCase = ["A long paragraph for summarization.", "Another paragraph for summarization."] _UpperCAmelCase = [0, 57, 3018, 70307, 91, 2] _UpperCAmelCase = self.tokenizer( a_ , max_length=len(a_ ) , padding=a_ , truncation=a_ , return_tensors="pt" ) self.assertIsInstance(a_ , a_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) _UpperCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(a_ , a_ ) def _a ( self ) -> str: if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = "I was born in 92000, and this is falsé." _UpperCAmelCase = tokenizer.tokenize(a_ ) _UpperCAmelCase = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) _UpperCAmelCase = tokenizer.encode(a_ , add_special_tokens=a_ ) _UpperCAmelCase = rust_tokenizer.encode(a_ , add_special_tokens=a_ ) self.assertListEqual(a_ , a_ ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(a_ ) _UpperCAmelCase = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ , a_ ) @slow def _a ( self ) -> Dict: # fmt: off _UpperCAmelCase = {"input_ids": [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. _UpperCAmelCase = [ "Le transformeur est un modèle d'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="moussaKam/mbarthez" , revision="c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6" , sequences=a_ , )
657
1
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = len(UpperCamelCase__ ) _UpperCAmelCase = len(matrix[0] ) _UpperCAmelCase = min(UpperCamelCase__ , UpperCamelCase__ ) for row in range(UpperCamelCase__ ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , UpperCamelCase__ ): _UpperCAmelCase = matrix[col][row] / matrix[row][row] for i in range(UpperCamelCase__ , UpperCamelCase__ ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows _UpperCAmelCase = True for i in range(row + 1 , UpperCamelCase__ ): if matrix[i][row] != 0: _UpperCAmelCase , _UpperCAmelCase = matrix[i], matrix[row] _UpperCAmelCase = False break if reduce: rank -= 1 for i in range(UpperCamelCase__ ): _UpperCAmelCase = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): _UpperCAmelCase = f"Input value of [number={number}] must be an integer" raise TypeError(UpperCamelCase__ ) if number < 0: return False _UpperCAmelCase = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
657
1
"""simple docstring""" from PIL import Image def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" def brightness(UpperCamelCase__ ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError("level must be between -255.0 (black) and 255.0 (white)" ) return img.point(UpperCamelCase__ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change brightness to 100 __magic_name__ = change_brightness(img, 1_00) brigt_img.save('''image_data/lena_brightness.png''', format='''png''')
657
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __magic_name__ = logging.get_logger(__name__) __magic_name__ = Dict[str, Any] __magic_name__ = List[Prediction] @add_end_docstrings(lowerCamelCase ) class _lowerCAmelCase ( lowerCamelCase ): def __init__( self , *a_ , **a_ ) -> Optional[int]: super().__init__(*a_ , **a_ ) if self.framework == "tf": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) requires_backends(self , "vision" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _a ( self , **a_ ) -> List[str]: _UpperCAmelCase = {} if "threshold" in kwargs: _UpperCAmelCase = kwargs["threshold"] return {}, {}, postprocess_kwargs def __call__( self , *a_ , **a_ ) -> Union[Predictions, List[Prediction]]: return super().__call__(*a_ , **a_ ) def _a ( self , a_ ) -> Optional[Any]: _UpperCAmelCase = load_image(a_ ) _UpperCAmelCase = torch.IntTensor([[image.height, image.width]] ) _UpperCAmelCase = self.image_processor(images=[image] , return_tensors="pt" ) if self.tokenizer is not None: _UpperCAmelCase = self.tokenizer(text=inputs["words"] , boxes=inputs["boxes"] , return_tensors="pt" ) _UpperCAmelCase = target_size return inputs def _a ( self , a_ ) -> Optional[Any]: _UpperCAmelCase = model_inputs.pop("target_size" ) _UpperCAmelCase = self.model(**a_ ) _UpperCAmelCase = outputs.__class__({"target_size": target_size, **outputs} ) if self.tokenizer is not None: _UpperCAmelCase = model_inputs["bbox"] return model_outputs def _a ( self , a_ , a_=0.9 ) -> int: _UpperCAmelCase = model_outputs["target_size"] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. _UpperCAmelCase , _UpperCAmelCase = target_size[0].tolist() def unnormalize(a_ ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) _UpperCAmelCase , _UpperCAmelCase = model_outputs["logits"].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) _UpperCAmelCase = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] _UpperCAmelCase = [unnormalize(a_ ) for bbox in model_outputs["bbox"].squeeze(0 )] _UpperCAmelCase = ["score", "label", "box"] _UpperCAmelCase = [dict(zip(a_ , a_ ) ) for vals in zip(scores.tolist() , a_ , a_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel _UpperCAmelCase = self.image_processor.post_process_object_detection(a_ , a_ , a_ ) _UpperCAmelCase = raw_annotations[0] _UpperCAmelCase = raw_annotation["scores"] _UpperCAmelCase = raw_annotation["labels"] _UpperCAmelCase = raw_annotation["boxes"] _UpperCAmelCase = scores.tolist() _UpperCAmelCase = [self.model.config.idalabel[label.item()] for label in labels] _UpperCAmelCase = [self._get_bounding_box(a_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] _UpperCAmelCase = ["score", "label", "box"] _UpperCAmelCase = [ dict(zip(a_ , a_ ) ) for vals in zip(raw_annotation["scores"] , raw_annotation["labels"] , raw_annotation["boxes"] ) ] return annotation def _a ( self , a_ ) -> Dict[str, int]: if self.framework != "pt": raise ValueError("The ObjectDetectionPipeline is only available in PyTorch." ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = box.int().tolist() _UpperCAmelCase = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
657
1
"""simple docstring""" import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Dict = '''encodec''' def __init__( self , a_=[1.5, 3.0, 6.0, 12.0, 24.0] , a_=24000 , a_=1 , a_=False , a_=None , a_=None , a_=128 , a_=32 , a_=1 , a_=[8, 5, 4, 2] , a_="weight_norm" , a_=7 , a_=7 , a_=3 , a_=2 , a_=True , a_="reflect" , a_=2 , a_=2 , a_=1.0 , a_=1024 , a_=None , a_=True , **a_ , ) -> Union[str, Any]: _UpperCAmelCase = target_bandwidths _UpperCAmelCase = sampling_rate _UpperCAmelCase = audio_channels _UpperCAmelCase = normalize _UpperCAmelCase = chunk_length_s _UpperCAmelCase = overlap _UpperCAmelCase = hidden_size _UpperCAmelCase = num_filters _UpperCAmelCase = num_residual_layers _UpperCAmelCase = upsampling_ratios _UpperCAmelCase = norm_type _UpperCAmelCase = kernel_size _UpperCAmelCase = last_kernel_size _UpperCAmelCase = residual_kernel_size _UpperCAmelCase = dilation_growth_rate _UpperCAmelCase = use_causal_conv _UpperCAmelCase = pad_mode _UpperCAmelCase = compress _UpperCAmelCase = num_lstm_layers _UpperCAmelCase = trim_right_ratio _UpperCAmelCase = codebook_size _UpperCAmelCase = codebook_dim if codebook_dim is not None else hidden_size _UpperCAmelCase = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f"self.norm_type must be one of `\"weight_norm\"`, `\"time_group_norm\"`), got {self.norm_type}" ) super().__init__(**a_ ) @property def _a ( self ) -> Optional[int]: if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _a ( self ) -> Optional[int]: if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _a ( self ) -> int: _UpperCAmelCase = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _a ( self ) -> int: return int(1000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" def merge(UpperCamelCase__ , UpperCamelCase__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(UpperCamelCase__ ) <= 1: return collection _UpperCAmelCase = len(UpperCamelCase__ ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() __magic_name__ = input('''Enter numbers separated by a comma:\n''').strip() __magic_name__ = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
657
1
"""simple docstring""" import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=False ): """simple docstring""" try: _UpperCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _UpperCAmelCase = default else: # KEY is set, convert it to True or False. try: _UpperCAmelCase = strtobool(UpperCamelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"If set, {key} must be yes or no." ) return _value __magic_name__ = parse_flag_from_env('''RUN_SLOW''', default=False) __magic_name__ = parse_flag_from_env('''RUN_REMOTE''', default=False) __magic_name__ = parse_flag_from_env('''RUN_LOCAL''', default=True) __magic_name__ = parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression __magic_name__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') __magic_name__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') __magic_name__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio __magic_name__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam __magic_name__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility __magic_name__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows __magic_name__ = pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import faiss # noqa except ImportError: _UpperCAmelCase = unittest.skip("test requires faiss" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import regex # noqa except ImportError: _UpperCAmelCase = unittest.skip("test requires regex" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import elasticsearch # noqa except ImportError: _UpperCAmelCase = unittest.skip("test requires elasticsearch" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import sqlalchemy # noqa except ImportError: _UpperCAmelCase = unittest.skip("test requires sqlalchemy" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not config.TORCH_AVAILABLE: _UpperCAmelCase = unittest.skip("test requires PyTorch" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not config.TF_AVAILABLE: _UpperCAmelCase = unittest.skip("test requires TensorFlow" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not config.JAX_AVAILABLE: _UpperCAmelCase = unittest.skip("test requires JAX" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not config.PIL_AVAILABLE: _UpperCAmelCase = unittest.skip("test requires Pillow" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(UpperCamelCase__ ) else: return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(UpperCamelCase__ ) else: return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(UpperCamelCase__ ) else: return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" def _require_spacy_model(UpperCamelCase__ ): try: import spacy # noqa F401 spacy.load(UpperCamelCase__ ) except ImportError: return unittest.skip("test requires spacy" )(UpperCamelCase__ ) except OSError: return unittest.skip("test requires spacy model '{}'".format(UpperCamelCase__ ) )(UpperCamelCase__ ) else: return test_case return _require_spacy_model def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(UpperCamelCase__ ) else: return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(UpperCamelCase__ ) else: return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: _UpperCAmelCase = unittest.skip("test is slow" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not _run_local_tests or _run_local_tests == 0: _UpperCAmelCase = unittest.skip("test is local" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: _UpperCAmelCase = unittest.skip("test is packaged" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: _UpperCAmelCase = unittest.skip("test requires remote" )(UpperCamelCase__ ) return test_case def __lowerCamelCase ( *UpperCamelCase__ ): """simple docstring""" def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(UpperCamelCase__ ) and name.startswith("test" ): for decorator in decorators: _UpperCAmelCase = decorator(UpperCamelCase__ ) setattr(cls , UpperCamelCase__ , UpperCamelCase__ ) return cls return decorate class _lowerCAmelCase ( lowerCamelCase ): pass class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Optional[int] = 0 lowercase_ : str = 1 lowercase_ : Any = 2 @contextmanager def __lowerCamelCase ( UpperCamelCase__=OfflineSimulationMode.CONNECTION_FAILS , UpperCamelCase__=1E-16 ): """simple docstring""" _UpperCAmelCase = requests.Session().request def timeout_request(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ): # Change the url to an invalid url so that the connection hangs _UpperCAmelCase = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"Tried a call to {url} in offline mode with no timeout set. Please set a timeout." ) _UpperCAmelCase = timeout try: return online_request(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier _UpperCAmelCase = url _UpperCAmelCase = e.args[0] _UpperCAmelCase = (max_retry_error.args[0].replace("10.255.255.1" , f"OfflineMock[{url}]" ),) _UpperCAmelCase = (max_retry_error,) raise def raise_connection_error(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ): raise requests.ConnectionError("Offline mode is enabled." , request=UpperCamelCase__ ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send" , UpperCamelCase__ ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request" , UpperCamelCase__ ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE" , UpperCamelCase__ ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def __lowerCamelCase ( *UpperCamelCase__ , **UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*UpperCamelCase__ , **UpperCamelCase__ ) as tmp_dir: try: os.chdir(UpperCamelCase__ ) yield finally: os.chdir(UpperCamelCase__ ) @contextmanager def __lowerCamelCase ( ): """simple docstring""" import gc gc.collect() _UpperCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def __lowerCamelCase ( ): """simple docstring""" import gc gc.collect() _UpperCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" return deepcopy(UpperCamelCase__ ).integers(0 , 100 , 10 ).tolist() == deepcopy(UpperCamelCase__ ).integers(0 , 100 , 10 ).tolist() def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ): try: return func(*UpperCamelCase__ , **UpperCamelCase__ ) except HTTPError as err: if str(UpperCamelCase__ ).startswith("500" ) or str(UpperCamelCase__ ).startswith("502" ): pytest.xfail(str(UpperCamelCase__ ) ) raise err return decorator.decorator(_wrapper , UpperCamelCase__ ) class _lowerCAmelCase : def __init__( self , a_ , a_ , a_ ) -> Optional[Any]: _UpperCAmelCase = returncode _UpperCAmelCase = stdout _UpperCAmelCase = stderr async def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" while True: _UpperCAmelCase = await stream.readline() if line: callback(UpperCamelCase__ ) else: break async def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=False ): """simple docstring""" if echo: print("\nRunning: " , " ".join(UpperCamelCase__ ) ) _UpperCAmelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=UpperCamelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=UpperCamelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _UpperCAmelCase = [] _UpperCAmelCase = [] def tee(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__="" ): _UpperCAmelCase = line.decode("utf-8" ).rstrip() sink.append(UpperCamelCase__ ) if not quiet: print(UpperCamelCase__ , UpperCamelCase__ , file=UpperCamelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda UpperCamelCase__ : tee(UpperCamelCase__ , UpperCamelCase__ , sys.stdout , label="stdout:" ) ), _read_stream(p.stderr , lambda UpperCamelCase__ : tee(UpperCamelCase__ , UpperCamelCase__ , sys.stderr , label="stderr:" ) ), ] , timeout=UpperCamelCase__ , ) return _RunOutput(await p.wait() , UpperCamelCase__ , UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=180 , UpperCamelCase__=False , UpperCamelCase__=True ): """simple docstring""" _UpperCAmelCase = asyncio.get_event_loop() _UpperCAmelCase = loop.run_until_complete( _stream_subprocess(UpperCamelCase__ , env=UpperCamelCase__ , stdin=UpperCamelCase__ , timeout=UpperCamelCase__ , quiet=UpperCamelCase__ , echo=UpperCamelCase__ ) ) _UpperCAmelCase = " ".join(UpperCamelCase__ ) if result.returncode > 0: _UpperCAmelCase = "\n".join(result.stderr ) raise RuntimeError( f"'{cmd_str}' failed with returncode {result.returncode}\n\n" f"The combined stderr from workers follows:\n{stderr}" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"'{cmd_str}' produced no output." ) return result def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = os.environ.get("PYTEST_XDIST_WORKER" , "gw0" ) _UpperCAmelCase = re.sub(r"^gw" , "" , UpperCamelCase__ , 0 , re.M ) return int(UpperCamelCase__ ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = 2_9500 _UpperCAmelCase = pytest_xdist_worker_id() return port + uniq_delta
657
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class _lowerCAmelCase : def __init__( self , a_ , a_=13 , a_=7 , a_=True , a_=True , a_=True , a_=99 , a_=32 , a_=5 , a_=4 , a_=37 , a_="gelu" , a_=0.1 , a_=0.1 , a_=512 , a_=16 , a_=2 , a_=0.02 , a_=3 , a_=4 , a_=None , ) -> List[str]: _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = seq_length _UpperCAmelCase = is_training _UpperCAmelCase = use_token_type_ids _UpperCAmelCase = use_labels _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = type_vocab_size _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = initializer_range _UpperCAmelCase = num_labels _UpperCAmelCase = num_choices _UpperCAmelCase = scope _UpperCAmelCase = self.vocab_size - 1 def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase = None if self.use_token_type_ids: _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) _UpperCAmelCase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) _UpperCAmelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def _a ( self , a_ , a_ , a_ , a_ , *a_ ) -> Optional[int]: _UpperCAmelCase = OpenAIGPTModel(config=a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model(a_ , token_type_ids=a_ , head_mask=a_ ) _UpperCAmelCase = model(a_ , token_type_ids=a_ ) _UpperCAmelCase = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self , a_ , a_ , a_ , a_ , *a_ ) -> List[Any]: _UpperCAmelCase = OpenAIGPTLMHeadModel(a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model(a_ , token_type_ids=a_ , labels=a_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self , a_ , a_ , a_ , a_ , *a_ ) -> Optional[Any]: _UpperCAmelCase = OpenAIGPTDoubleHeadsModel(a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model(a_ , token_type_ids=a_ , labels=a_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self , a_ , a_ , a_ , a_ , *a_ ) -> Dict: _UpperCAmelCase = self.num_labels _UpperCAmelCase = OpenAIGPTForSequenceClassification(a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = model(a_ , token_type_ids=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a ( self ) -> List[str]: _UpperCAmelCase = self.prepare_config_and_inputs() ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) = config_and_inputs _UpperCAmelCase = { "input_ids": input_ids, "token_type_ids": token_type_ids, "head_mask": head_mask, } return config, inputs_dict @require_torch class _lowerCAmelCase ( lowerCamelCase , lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase_ : Any = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) lowercase_ : Optional[Any] = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly lowercase_ : Union[str, Any] = ( { '''feature-extraction''': OpenAIGPTModel, '''text-classification''': OpenAIGPTForSequenceClassification, '''text-generation''': OpenAIGPTLMHeadModel, '''zero-shot''': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def _a ( self , a_ , a_ , a_ , a_ , a_ ) -> Any: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def _a ( self , a_ , a_ , a_=False ) -> Optional[int]: _UpperCAmelCase = super()._prepare_for_class(a_ , a_ , return_labels=a_ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": _UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=a_ , ) _UpperCAmelCase = inputs_dict["labels"] _UpperCAmelCase = inputs_dict["labels"] _UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=a_ , ) _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a_ ) return inputs_dict def _a ( self ) -> Optional[int]: _UpperCAmelCase = OpenAIGPTModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=a_ , n_embd=37 ) def _a ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def _a ( self ) -> Any: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*a_ ) def _a ( self ) -> Tuple: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*a_ ) def _a ( self ) -> List[Any]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*a_ ) def _a ( self ) -> List[str]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*a_ ) @slow def _a ( self ) -> int: for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = OpenAIGPTModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): @slow def _a ( self ) -> Any: _UpperCAmelCase = OpenAIGPTLMHeadModel.from_pretrained("openai-gpt" ) model.to(a_ ) _UpperCAmelCase = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=a_ ) # the president is _UpperCAmelCase = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 40477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the _UpperCAmelCase = model.generate(a_ , do_sample=a_ ) self.assertListEqual(output_ids[0].tolist() , a_ )
657
1
"""simple docstring""" from jiwer import compute_measures import datasets __magic_name__ = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __magic_name__ = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __magic_name__ = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): def _a ( self ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/jitsi/jiwer/"] , reference_urls=[ "https://en.wikipedia.org/wiki/Word_error_rate", ] , ) def _a ( self , a_=None , a_=None , a_=False ) -> List[str]: if concatenate_texts: return compute_measures(a_ , a_ )["wer"] else: _UpperCAmelCase = 0 _UpperCAmelCase = 0 for prediction, reference in zip(a_ , a_ ): _UpperCAmelCase = compute_measures(a_ , a_ ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
657
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=10 ): """simple docstring""" _UpperCAmelCase = [] for _ in range(UpperCamelCase__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=10 ): """simple docstring""" _UpperCAmelCase = [] for step in range(UpperCamelCase__ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase = os.path.join(UpperCamelCase__ , "schedule.bin" ) torch.save(scheduler.state_dict() , UpperCamelCase__ ) _UpperCAmelCase = torch.load(UpperCamelCase__ ) scheduler.load_state_dict(UpperCamelCase__ ) return lrs @require_torch class _lowerCAmelCase ( unittest.TestCase ): def _a ( self , a_ , a_ , a_ ) -> Optional[int]: self.assertEqual(len(a_ ) , len(a_ ) ) for a, b in zip(a_ , a_ ): self.assertAlmostEqual(a_ , a_ , delta=a_ ) def _a ( self ) -> str: _UpperCAmelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=a_ ) _UpperCAmelCase = torch.tensor([0.4, 0.2, -0.5] ) _UpperCAmelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _UpperCAmelCase = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(100 ): _UpperCAmelCase = criterion(a_ , a_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=a_ ) _UpperCAmelCase = torch.tensor([0.4, 0.2, -0.5] ) _UpperCAmelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _UpperCAmelCase = Adafactor( params=[w] , lr=1e-2 , eps=(1e-30, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=a_ , weight_decay=0.0 , relative_step=a_ , scale_parameter=a_ , warmup_init=a_ , ) for _ in range(1000 ): _UpperCAmelCase = criterion(a_ , a_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): lowercase_ : List[Any] = nn.Linear(50 , 50 ) if is_torch_available() else None lowercase_ : Tuple = AdamW(m.parameters() , lr=10.0 ) if is_torch_available() else None lowercase_ : Dict = 10 def _a ( self , a_ , a_ , a_ , a_=None ) -> Union[str, Any]: self.assertEqual(len(a_ ) , len(a_ ) ) for a, b in zip(a_ , a_ ): self.assertAlmostEqual(a_ , a_ , delta=a_ , msg=a_ ) def _a ( self ) -> List[Any]: _UpperCAmelCase = {"num_warmup_steps": 2, "num_training_steps": 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _UpperCAmelCase = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"num_warmup_steps": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, "num_cycles": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, "power": 2.0, "lr_end": 1e-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"num_warmup_steps": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): _UpperCAmelCase , _UpperCAmelCase = data _UpperCAmelCase = scheduler_func(self.optimizer , **a_ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _UpperCAmelCase = unwrap_schedule(a_ , self.num_steps ) self.assertListAlmostEqual( a_ , a_ , tol=1e-2 , msg=f"failed for {scheduler_func} in normal scheduler" , ) _UpperCAmelCase = scheduler_func(self.optimizer , **a_ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(a_ ) # wrap to test picklability of the schedule _UpperCAmelCase = unwrap_and_save_reload_schedule(a_ , self.num_steps ) self.assertListEqual(a_ , a_ , msg=f"failed for {scheduler_func} in save and reload" ) class _lowerCAmelCase : def __init__( self , a_ ) -> Union[str, Any]: _UpperCAmelCase = fn def __call__( self , *a_ , **a_ ) -> Union[str, Any]: return self.fn(*a_ , **a_ ) @classmethod def _a ( self , a_ ) -> Dict: _UpperCAmelCase = list(map(self , scheduler.lr_lambdas ) )
657
1
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class _lowerCAmelCase ( unittest.TestCase ): def _a ( self ) -> Optional[Any]: _UpperCAmelCase = ["a", "b", "c"] # Defaults to last layer if both are None _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , a_ , a_ ) self.assertEqual(a_ , ["c"] ) self.assertEqual(a_ , [2] ) # Out indices set to match out features _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(["a", "c"] , a_ , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [0, 2] ) # Out features set to match out indices _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , [0, 2] , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [0, 2] ) # Out features selected from negative indices _UpperCAmelCase , _UpperCAmelCase = get_aligned_output_features_output_indices(a_ , [-3, -1] , a_ ) self.assertEqual(a_ , ["a", "c"] ) self.assertEqual(a_ , [-3, -1] ) def _a ( self ) -> Optional[int]: # Stage names must be set with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , a_ ) # Out features must be a list with self.assertRaises(a_ ): verify_out_features_out_indices(("a", "b") , (0, 1) , ["a", "b"] ) # Out features must be a subset of stage names with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , ["a"] ) # Out indices must be a list or tuple with self.assertRaises(a_ ): verify_out_features_out_indices(a_ , 0 , ["a", "b"] ) # Out indices must be a subset of stage names with self.assertRaises(a_ ): verify_out_features_out_indices(a_ , (0, 1) , ["a"] ) # Out features and out indices must be the same length with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0,) , ["a", "b", "c"] ) # Out features should match out indices with self.assertRaises(a_ ): verify_out_features_out_indices(["a", "b"] , (0, 2) , ["a", "b", "c"] ) # Out features and out indices should be in order with self.assertRaises(a_ ): verify_out_features_out_indices(["b", "a"] , (0, 1) , ["a", "b"] ) # Check passes with valid inputs verify_out_features_out_indices(["a", "b", "d"] , (0, 1, -1) , ["a", "b", "c", "d"] ) def _a ( self ) -> int: _UpperCAmelCase = BackboneMixin() _UpperCAmelCase = ["a", "b", "c"] _UpperCAmelCase = ["a", "c"] _UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly _UpperCAmelCase = ["a", "b"] self.assertEqual(backbone.out_features , ["a", "b"] ) self.assertEqual(backbone.out_indices , [0, 1] ) _UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [-3, -1] )
657
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __lowerCamelCase ( UpperCamelCase__=None ): """simple docstring""" if subparsers is not None: _UpperCAmelCase = subparsers.add_parser("test" ) else: _UpperCAmelCase = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=UpperCamelCase__ , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase__ ) return parser def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: _UpperCAmelCase = script_name else: _UpperCAmelCase = f"--config_file={args.config_file} {script_name}" _UpperCAmelCase = ["accelerate-launch"] + test_args.split() _UpperCAmelCase = execute_subprocess_async(UpperCamelCase__ , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = test_command_parser() _UpperCAmelCase = parser.parse_args() test_command(UpperCamelCase__ ) if __name__ == "__main__": main()
657
1
"""simple docstring""" class _lowerCAmelCase : def __init__( self , a_ = "" , a_ = False ) -> None: # Mapping from the first character of the prefix of the node _UpperCAmelCase = {} # A node will be a leaf if the tree contains its word _UpperCAmelCase = is_leaf _UpperCAmelCase = prefix def _a ( self , a_ ) -> tuple[str, str, str]: _UpperCAmelCase = 0 for q, w in zip(self.prefix , a_ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def _a ( self , a_ ) -> None: for word in words: self.insert(a_ ) def _a ( self , a_ ) -> None: # Case 1: If the word is the prefix of the node # Solution: We set the current node as leaf if self.prefix == word: _UpperCAmelCase = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: _UpperCAmelCase = RadixNode(prefix=a_ , is_leaf=a_ ) else: _UpperCAmelCase = self.nodes[word[0]] _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( a_ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(a_ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: _UpperCAmelCase = remaining_prefix _UpperCAmelCase = self.nodes[matching_string[0]] _UpperCAmelCase = RadixNode(a_ , a_ ) _UpperCAmelCase = aux_node if remaining_word == "": _UpperCAmelCase = True else: self.nodes[matching_string[0]].insert(a_ ) def _a ( self , a_ ) -> bool: _UpperCAmelCase = self.nodes.get(word[0] , a_ ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( a_ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(a_ ) def _a ( self , a_ ) -> bool: _UpperCAmelCase = self.nodes.get(word[0] , a_ ) if not incoming_node: return False else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = incoming_node.match( a_ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(a_ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: _UpperCAmelCase = list(self.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf self.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: _UpperCAmelCase = False # If there is 1 edge, we merge it with its child else: _UpperCAmelCase = list(incoming_node.nodes.values() )[0] _UpperCAmelCase = merging_node.is_leaf incoming_node.prefix += merging_node.prefix _UpperCAmelCase = merging_node.nodes return True def _a ( self , a_ = 0 ) -> None: if self.prefix != "": print("-" * height , self.prefix , " (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = "banana bananas bandana band apple all beast".split() _UpperCAmelCase = RadixNode() root.insert_many(UpperCamelCase__ ) assert all(root.find(UpperCamelCase__ ) for word in words ) assert not root.find("bandanas" ) assert not root.find("apps" ) root.delete("all" ) assert not root.find("all" ) root.delete("banana" ) assert not root.find("banana" ) assert root.find("bananas" ) return True def __lowerCamelCase ( ): """simple docstring""" assert test_trie() def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = RadixNode() _UpperCAmelCase = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(UpperCamelCase__ ) print("Words:" , UpperCamelCase__ ) print("Tree:" ) root.print_tree() if __name__ == "__main__": main()
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" return 10 - x * x def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) >= 0: raise ValueError("Wrong space!" ) _UpperCAmelCase = a while (b - a) >= 0.01: # Find middle point _UpperCAmelCase = (a + b) / 2 # Check if middle point is root if equation(UpperCamelCase__ ) == 0.0: break # Decide the side to repeat the steps if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) < 0: _UpperCAmelCase = c else: _UpperCAmelCase = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
657
1
"""simple docstring""" import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( lowerCamelCase , unittest.TestCase ): lowercase_ : Optional[Any] = TransfoXLTokenizer lowercase_ : List[str] = False lowercase_ : Tuple = False def _a ( self ) -> List[str]: super().setUp() _UpperCAmelCase = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def _a ( self , **a_ ) -> Dict: _UpperCAmelCase = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **a_ ) def _a ( self , a_ ) -> List[str]: _UpperCAmelCase = "<unk> UNwanted , running" _UpperCAmelCase = "<unk> unwanted, running" return input_text, output_text def _a ( self ) -> Any: _UpperCAmelCase = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=a_ ) _UpperCAmelCase = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(a_ , ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [0, 4, 8, 7] ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = TransfoXLTokenizer(lower_case=a_ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ) , ["hello", "!", "how", "are", "you", "?"] ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = TransfoXLTokenizer(lower_case=a_ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def _a ( self ) -> Tuple: _UpperCAmelCase = TransfoXLTokenizer(lower_case=a_ ) _UpperCAmelCase = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" _UpperCAmelCase = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(a_ ) , a_ ) self.assertEqual(tokenizer.convert_tokens_to_string(a_ ) , a_ ) def _a ( self ) -> Any: _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = len(a_ ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(a_ ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , "new1" )
657
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _lowerCAmelCase ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): lowercase_ : Tuple = [r'''h\.\d+\.attn\.bias''', r'''h\.\d+\.attn\.masked_bias'''] @register_to_config def __init__( self , a_ , a_ , a_ = None , a_ = 50257 , a_ = 1024 , a_ = 768 , a_ = 12 , a_ = 12 , a_ = None , a_ = "gelu_new" , a_ = 0.1 , a_ = 0.1 , a_ = 0.1 , a_ = 1e-5 , a_ = 0.02 , a_ = True , a_ = True , a_ = False , a_ = False , ) -> List[str]: super().__init__() _UpperCAmelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f"`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and" f" `n_embd`: {n_embd} are not equal." ) _UpperCAmelCase = prefix_inner_dim _UpperCAmelCase = prefix_hidden_dim _UpperCAmelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _UpperCAmelCase = ( nn.Linear(self.prefix_hidden_dim , a_ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _UpperCAmelCase = GPTaConfig( vocab_size=a_ , n_positions=a_ , n_embd=a_ , n_layer=a_ , n_head=a_ , n_inner=a_ , activation_function=a_ , resid_pdrop=a_ , embd_pdrop=a_ , attn_pdrop=a_ , layer_norm_epsilon=a_ , initializer_range=a_ , scale_attn_weights=a_ , use_cache=a_ , scale_attn_by_inverse_layer_idx=a_ , reorder_and_upcast_attn=a_ , ) _UpperCAmelCase = GPTaLMHeadModel(a_ ) def _a ( self , a_ , a_ , a_ = None , a_ = None , ) -> Tuple: _UpperCAmelCase = self.transformer.transformer.wte(a_ ) _UpperCAmelCase = self.encode_prefix(a_ ) _UpperCAmelCase = self.decode_prefix(a_ ) _UpperCAmelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _UpperCAmelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _UpperCAmelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _UpperCAmelCase = self.transformer(inputs_embeds=a_ , labels=a_ , attention_mask=a_ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _a ( self , a_ , a_ ) -> torch.Tensor: return torch.zeros(a_ , self.prefix_length , dtype=torch.intaa , device=a_ ) def _a ( self , a_ ) -> Union[str, Any]: return self.encode_prefix(a_ ) @torch.no_grad() def _a ( self , a_ , a_ , a_ ) -> Union[str, Any]: _UpperCAmelCase = torch.split(a_ , 1 , dim=0 ) _UpperCAmelCase = [] _UpperCAmelCase = [] for feature in features: _UpperCAmelCase = self.decode_prefix(feature.to(a_ ) ) # back to the clip feature # Only support beam search for now _UpperCAmelCase , _UpperCAmelCase = self.generate_beam( input_embeds=a_ , device=a_ , eos_token_id=a_ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _UpperCAmelCase = torch.stack(a_ ) _UpperCAmelCase = torch.stack(a_ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _a ( self , a_=None , a_=None , a_=None , a_ = 5 , a_ = 67 , a_ = 1.0 , a_ = None , ) -> Optional[Any]: _UpperCAmelCase = eos_token_id _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = torch.ones(a_ , device=a_ , dtype=torch.int ) _UpperCAmelCase = torch.zeros(a_ , device=a_ , dtype=torch.bool ) if input_embeds is not None: _UpperCAmelCase = input_embeds else: _UpperCAmelCase = self.transformer.transformer.wte(a_ ) for i in range(a_ ): _UpperCAmelCase = self.transformer(inputs_embeds=a_ ) _UpperCAmelCase = outputs.logits _UpperCAmelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _UpperCAmelCase = logits.softmax(-1 ).log() if scores is None: _UpperCAmelCase , _UpperCAmelCase = logits.topk(a_ , -1 ) _UpperCAmelCase = generated.expand(a_ , *generated.shape[1:] ) _UpperCAmelCase , _UpperCAmelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _UpperCAmelCase = next_tokens else: _UpperCAmelCase = tokens.expand(a_ , *tokens.shape[1:] ) _UpperCAmelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _UpperCAmelCase = -float(np.inf ) _UpperCAmelCase = 0 _UpperCAmelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _UpperCAmelCase = scores_sum / seq_lengths[:, None] _UpperCAmelCase , _UpperCAmelCase = scores_sum_average.view(-1 ).topk(a_ , -1 ) _UpperCAmelCase = next_tokens // scores_sum.shape[1] _UpperCAmelCase = seq_lengths[next_tokens_source] _UpperCAmelCase = next_tokens % scores_sum.shape[1] _UpperCAmelCase = next_tokens.unsqueeze(1 ) _UpperCAmelCase = tokens[next_tokens_source] _UpperCAmelCase = torch.cat((tokens, next_tokens) , dim=1 ) _UpperCAmelCase = generated[next_tokens_source] _UpperCAmelCase = scores_sum_average * seq_lengths _UpperCAmelCase = is_stopped[next_tokens_source] _UpperCAmelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _UpperCAmelCase = torch.cat((generated, next_token_embed) , dim=1 ) _UpperCAmelCase = is_stopped + next_tokens.eq(a_ ).squeeze() if is_stopped.all(): break _UpperCAmelCase = scores / seq_lengths _UpperCAmelCase = scores.argsort(descending=a_ ) # tokens tensors are already padded to max_seq_length _UpperCAmelCase = [tokens[i] for i in order] _UpperCAmelCase = torch.stack(a_ , dim=0 ) _UpperCAmelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
657
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json''' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : str = '''trocr''' lowercase_ : Optional[int] = ['''past_key_values'''] lowercase_ : Any = { '''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model''', '''num_hidden_layers''': '''decoder_layers''', } def __init__( self , a_=50265 , a_=1024 , a_=12 , a_=16 , a_=4096 , a_="gelu" , a_=512 , a_=0.1 , a_=0.0 , a_=0.0 , a_=2 , a_=0.02 , a_=0.0 , a_=True , a_=False , a_=True , a_=True , a_=1 , a_=0 , a_=2 , **a_ , ) -> Any: _UpperCAmelCase = vocab_size _UpperCAmelCase = d_model _UpperCAmelCase = decoder_layers _UpperCAmelCase = decoder_attention_heads _UpperCAmelCase = decoder_ffn_dim _UpperCAmelCase = activation_function _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = activation_dropout _UpperCAmelCase = init_std _UpperCAmelCase = decoder_layerdrop _UpperCAmelCase = use_cache _UpperCAmelCase = scale_embedding _UpperCAmelCase = use_learned_position_embeddings _UpperCAmelCase = layernorm_embedding super().__init__( pad_token_id=a_ , bos_token_id=a_ , eos_token_id=a_ , decoder_start_token_id=a_ , **a_ , )
657
"""simple docstring""" from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable __magic_name__ = {'''configuration_gpt_neox''': ['''GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXConfig''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = ['''GPTNeoXTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXForCausalLM''', '''GPTNeoXForQuestionAnswering''', '''GPTNeoXForSequenceClassification''', '''GPTNeoXForTokenClassification''', '''GPTNeoXLayer''', '''GPTNeoXModel''', '''GPTNeoXPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
657
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _lowerCAmelCase ( lowerCamelCase , unittest.TestCase ): lowercase_ : Union[str, Any] = KandinskyVaaPipeline lowercase_ : Tuple = [ '''image_embeds''', '''negative_image_embeds''', ] lowercase_ : Tuple = ['''image_embeds''', '''negative_image_embeds'''] lowercase_ : Optional[Any] = [ '''generator''', '''height''', '''width''', '''latents''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] lowercase_ : Tuple = False @property def _a ( self ) -> List[Any]: return 32 @property def _a ( self ) -> List[Any]: return 32 @property def _a ( self ) -> List[str]: return self.time_input_dim @property def _a ( self ) -> Optional[Any]: return self.time_input_dim * 4 @property def _a ( self ) -> List[Any]: return 100 @property def _a ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCAmelCase = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } _UpperCAmelCase = UNetaDConditionModel(**a_ ) return model @property def _a ( self ) -> int: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def _a ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCAmelCase = VQModel(**self.dummy_movq_kwargs ) return model def _a ( self ) -> Dict: _UpperCAmelCase = self.dummy_unet _UpperCAmelCase = self.dummy_movq _UpperCAmelCase = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.00085 , beta_end=0.012 , clip_sample=a_ , set_alpha_to_one=a_ , steps_offset=1 , prediction_type="epsilon" , thresholding=a_ , ) _UpperCAmelCase = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def _a ( self , a_ , a_=0 ) -> Union[str, Any]: _UpperCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(a_ ) ).to(a_ ) _UpperCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( a_ ) if str(a_ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(a_ ) else: _UpperCAmelCase = torch.Generator(device=a_ ).manual_seed(a_ ) _UpperCAmelCase = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def _a ( self ) -> str: _UpperCAmelCase = "cpu" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = self.pipeline_class(**a_ ) _UpperCAmelCase = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase = pipe(**self.get_dummy_inputs(a_ ) ) _UpperCAmelCase = output.images _UpperCAmelCase = pipe( **self.get_dummy_inputs(a_ ) , return_dict=a_ , )[0] _UpperCAmelCase = image[0, -3:, -3:, -1] _UpperCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase = np.array( [0.6237976, 1.0, 0.36441332, 1.0, 0.70639634, 0.29877186, 0.85652125, 0.5216843, 0.54454046] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): def _a ( self ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self ) -> Optional[Any]: _UpperCAmelCase = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) _UpperCAmelCase = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(a_ ) _UpperCAmelCase = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) _UpperCAmelCase = pipeline.to(a_ ) pipeline.set_progress_bar_config(disable=a_ ) _UpperCAmelCase = "red cat, 4k photo" _UpperCAmelCase = torch.Generator(device="cuda" ).manual_seed(0 ) _UpperCAmelCase , _UpperCAmelCase = pipe_prior( a_ , generator=a_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() _UpperCAmelCase = torch.Generator(device="cuda" ).manual_seed(0 ) _UpperCAmelCase = pipeline( image_embeds=a_ , negative_image_embeds=a_ , generator=a_ , num_inference_steps=100 , output_type="np" , ) _UpperCAmelCase = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(a_ , a_ )
657
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json''', # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Union[str, Any] = '''convbert''' def __init__( self , a_=30522 , a_=768 , a_=12 , a_=12 , a_=3072 , a_="gelu" , a_=0.1 , a_=0.1 , a_=512 , a_=2 , a_=0.02 , a_=1e-12 , a_=1 , a_=0 , a_=2 , a_=768 , a_=2 , a_=9 , a_=1 , a_=None , **a_ , ) -> Tuple: super().__init__( pad_token_id=a_ , bos_token_id=a_ , eos_token_id=a_ , **a_ , ) _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = type_vocab_size _UpperCAmelCase = initializer_range _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = embedding_size _UpperCAmelCase = head_ratio _UpperCAmelCase = conv_kernel_size _UpperCAmelCase = num_groups _UpperCAmelCase = classifier_dropout class _lowerCAmelCase ( lowerCamelCase ): @property def _a ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCAmelCase = {0: "batch", 1: "choice", 2: "sequence"} else: _UpperCAmelCase = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
657
1
"""simple docstring""" import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class _lowerCAmelCase : def __init__( self , a_ , a_=13 , a_=10 , a_=3 , a_=2 , a_=2 , a_=2 , a_=True , a_=True , a_=32 , a_=5 , a_=4 , a_=37 , a_="gelu" , a_=0.1 , a_=0.1 , a_=10 , a_=0.02 , a_=0.9 , a_=None , ) -> List[str]: _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = patch_size _UpperCAmelCase = tubelet_size _UpperCAmelCase = num_frames _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = initializer_range _UpperCAmelCase = mask_ratio _UpperCAmelCase = scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame _UpperCAmelCase = (image_size // patch_size) ** 2 _UpperCAmelCase = (num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos _UpperCAmelCase = int(mask_ratio * self.seq_length ) def _a ( self ) -> Optional[Any]: _UpperCAmelCase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def _a ( self ) -> List[Any]: return VideoMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , tubelet_size=self.tubelet_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=a_ , initializer_range=self.initializer_range , ) def _a ( self , a_ , a_ , a_ ) -> Tuple: _UpperCAmelCase = VideoMAEModel(config=a_ ) model.to(a_ ) model.eval() _UpperCAmelCase = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self , a_ , a_ , a_ ) -> Optional[int]: _UpperCAmelCase = VideoMAEForPreTraining(a_ ) model.to(a_ ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch _UpperCAmelCase = torch.ones((self.num_masks,) ) _UpperCAmelCase = torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) _UpperCAmelCase = mask.expand(self.batch_size , -1 ).bool() _UpperCAmelCase = model(a_ , a_ ) # model only returns predictions for masked patches _UpperCAmelCase = mask.sum().item() _UpperCAmelCase = 3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_masked_patches, decoder_num_labels) ) def _a ( self ) -> int: _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase_ : int = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) lowercase_ : Union[str, Any] = ( {'''feature-extraction''': VideoMAEModel, '''video-classification''': VideoMAEForVideoClassification} if is_torch_available() else {} ) lowercase_ : Tuple = False lowercase_ : int = False lowercase_ : str = False lowercase_ : Dict = False def _a ( self ) -> Dict: _UpperCAmelCase = VideoMAEModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=a_ , has_text_modality=a_ , hidden_size=37 ) def _a ( self , a_ , a_ , a_=False ) -> Dict: _UpperCAmelCase = copy.deepcopy(a_ ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch _UpperCAmelCase = torch.ones((self.model_tester.num_masks,) ) _UpperCAmelCase = torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) _UpperCAmelCase = mask.expand(self.model_tester.batch_size , -1 ).bool() _UpperCAmelCase = bool_masked_pos.to(a_ ) if return_labels: if model_class in [ *get_values(a_ ), ]: _UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a_ ) return inputs_dict def _a ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason="VideoMAE does not use inputs_embeds" ) def _a ( self ) -> List[str]: pass def _a ( self ) -> Optional[int]: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def _a ( self ) -> List[Any]: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(a_ ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , a_ ) def _a ( self ) -> int: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def _a ( self ) -> Optional[Any]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*a_ ) @slow def _a ( self ) -> Optional[int]: for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = VideoMAEModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def _a ( self ) -> Any: if not self.has_attentions: pass else: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase = True for model_class in self.all_model_classes: _UpperCAmelCase = self.model_tester.seq_length - self.model_tester.num_masks _UpperCAmelCase = ( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) _UpperCAmelCase = True _UpperCAmelCase = False _UpperCAmelCase = True _UpperCAmelCase = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(a_ , a_ ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(a_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _UpperCAmelCase = True _UpperCAmelCase = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(a_ , a_ ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(a_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) _UpperCAmelCase = len(a_ ) # Check attention is always last and order is fine _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(a_ , a_ ) ) self.assertEqual(out_len + 1 , len(a_ ) ) _UpperCAmelCase = outputs.attentions self.assertEqual(len(a_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def _a ( self ) -> Dict: def check_hidden_states_output(a_ , a_ , a_ ): _UpperCAmelCase = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(a_ , a_ ) ) _UpperCAmelCase = outputs.hidden_states _UpperCAmelCase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(a_ ) , a_ ) _UpperCAmelCase = self.model_tester.seq_length - self.model_tester.num_masks _UpperCAmelCase = num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = True check_hidden_states_output(a_ , a_ , a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(a_ , a_ , a_ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _a ( self ) -> str: pass def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename="eating_spaghetti.npy" , repo_type="dataset" ) _UpperCAmelCase = np.load(UpperCamelCase__ ) return list(UpperCamelCase__ ) @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): @cached_property def _a ( self ) -> Tuple: # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def _a ( self ) -> List[str]: _UpperCAmelCase = VideoMAEForVideoClassification.from_pretrained("MCG-NJU/videomae-base-finetuned-kinetics" ).to( a_ ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_video() _UpperCAmelCase = image_processor(a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): _UpperCAmelCase = model(**a_ ) # verify the logits _UpperCAmelCase = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape , a_ ) _UpperCAmelCase = torch.tensor([0.3669, -0.0688, -0.2421] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1e-4 ) ) @slow def _a ( self ) -> str: _UpperCAmelCase = VideoMAEForPreTraining.from_pretrained("MCG-NJU/videomae-base-short" ).to(a_ ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_video() _UpperCAmelCase = image_processor(a_ , return_tensors="pt" ).to(a_ ) # add boolean mask, indicating which patches to mask _UpperCAmelCase = hf_hub_download(repo_id="hf-internal-testing/bool-masked-pos" , filename="bool_masked_pos.pt" ) _UpperCAmelCase = torch.load(a_ ) # forward pass with torch.no_grad(): _UpperCAmelCase = model(**a_ ) # verify the logits _UpperCAmelCase = torch.Size([1, 1408, 1536] ) _UpperCAmelCase = torch.tensor( [[0.7994, 0.9612, 0.8508], [0.7401, 0.8958, 0.8302], [0.5862, 0.7468, 0.7325]] , device=a_ ) self.assertEqual(outputs.logits.shape , a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , a_ , atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) _UpperCAmelCase = torch.tensor([0.5142] , device=a_ ) self.assertTrue(torch.allclose(outputs.loss , a_ , atol=1e-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) _UpperCAmelCase = VideoMAEForPreTraining.from_pretrained("MCG-NJU/videomae-base-short" , norm_pix_loss=a_ ).to( a_ ) with torch.no_grad(): _UpperCAmelCase = model(**a_ ) _UpperCAmelCase = torch.tensor(torch.tensor([0.6469] ) , device=a_ ) self.assertTrue(torch.allclose(outputs.loss , a_ , atol=1e-4 ) )
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" return "".join([hex(UpperCamelCase__ )[2:].zfill(2 ).upper() for byte in list(UpperCamelCase__ )] ) def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" if (len(UpperCamelCase__ ) % 2) != 0: raise ValueError( "Base16 encoded data is invalid:\nData does not have an even number of hex digits." ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(UpperCamelCase__ ) <= set("0123456789ABCDEF" ): raise ValueError( "Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters." ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(UpperCamelCase__ ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
657
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
657
"""simple docstring""" def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" try: _UpperCAmelCase = float(UpperCamelCase__ ) except ValueError: raise ValueError("Please enter a valid number" ) _UpperCAmelCase = decimal - int(UpperCamelCase__ ) if fractional_part == 0: return int(UpperCamelCase__ ), 1 else: _UpperCAmelCase = len(str(UpperCamelCase__ ).split("." )[1] ) _UpperCAmelCase = int(decimal * (10**number_of_frac_digits) ) _UpperCAmelCase = 10**number_of_frac_digits _UpperCAmelCase , _UpperCAmelCase = denominator, numerator while True: _UpperCAmelCase = dividend % divisor if remainder == 0: break _UpperCAmelCase , _UpperCAmelCase = divisor, remainder _UpperCAmelCase , _UpperCAmelCase = numerator / divisor, denominator / divisor return int(UpperCamelCase__ ), int(UpperCamelCase__ ) if __name__ == "__main__": print(f'''{decimal_to_fraction(2) = }''') print(f'''{decimal_to_fraction(89.0) = }''') print(f'''{decimal_to_fraction("67") = }''') print(f'''{decimal_to_fraction("45.0") = }''') print(f'''{decimal_to_fraction(1.5) = }''') print(f'''{decimal_to_fraction("6.25") = }''') print(f'''{decimal_to_fraction("78td") = }''')
657
1
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __magic_name__ = True except (ImportError, AttributeError): __magic_name__ = object def __lowerCamelCase ( *UpperCamelCase__ , **UpperCamelCase__ ): """simple docstring""" pass __magic_name__ = False __magic_name__ = logging.get_logger('''transformers-cli/serving''') def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(UpperCamelCase__ , args.host , args.port , args.workers ) class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : dict class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : List[str] lowercase_ : Optional[List[int]] class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : str class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Any class _lowerCAmelCase ( lowerCamelCase ): @staticmethod def _a ( a_ ) -> Union[str, Any]: _UpperCAmelCase = parser.add_parser( "serve" , help="CLI tool to run inference requests through REST and GraphQL endpoints." ) serve_parser.add_argument( "--task" , type=a_ , choices=get_supported_tasks() , help="The task to run the pipeline on" , ) serve_parser.add_argument("--host" , type=a_ , default="localhost" , help="Interface the server will listen on." ) serve_parser.add_argument("--port" , type=a_ , default=8888 , help="Port the serving will listen to." ) serve_parser.add_argument("--workers" , type=a_ , default=1 , help="Number of http workers" ) serve_parser.add_argument("--model" , type=a_ , help="Model's name or path to stored model." ) serve_parser.add_argument("--config" , type=a_ , help="Model's config name or path to stored model." ) serve_parser.add_argument("--tokenizer" , type=a_ , help="Tokenizer name to use." ) serve_parser.add_argument( "--device" , type=a_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) serve_parser.set_defaults(func=a_ ) def __init__( self , a_ , a_ , a_ , a_ ) -> Tuple: _UpperCAmelCase = pipeline _UpperCAmelCase = host _UpperCAmelCase = port _UpperCAmelCase = workers if not _serve_dependencies_installed: raise RuntimeError( "Using serve command requires FastAPI and uvicorn. " "Please install transformers with [serving]: pip install \"transformers[serving]\"." "Or install FastAPI and uvicorn separately." ) else: logger.info(f"Serving model over {host}:{port}" ) _UpperCAmelCase = FastAPI( routes=[ APIRoute( "/" , self.model_info , response_model=a_ , response_class=a_ , methods=["GET"] , ), APIRoute( "/tokenize" , self.tokenize , response_model=a_ , response_class=a_ , methods=["POST"] , ), APIRoute( "/detokenize" , self.detokenize , response_model=a_ , response_class=a_ , methods=["POST"] , ), APIRoute( "/forward" , self.forward , response_model=a_ , response_class=a_ , methods=["POST"] , ), ] , timeout=600 , ) def _a ( self ) -> str: run(self._app , host=self.host , port=self.port , workers=self.workers ) def _a ( self ) -> List[str]: return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def _a ( self , a_ = Body(a_ , embed=a_ ) , a_ = Body(a_ , embed=a_ ) ) -> Union[str, Any]: try: _UpperCAmelCase = self._pipeline.tokenizer.tokenize(a_ ) if return_ids: _UpperCAmelCase = self._pipeline.tokenizer.convert_tokens_to_ids(a_ ) return ServeTokenizeResult(tokens=a_ , tokens_ids=a_ ) else: return ServeTokenizeResult(tokens=a_ ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(a_ )} ) def _a ( self , a_ = Body(a_ , embed=a_ ) , a_ = Body(a_ , embed=a_ ) , a_ = Body(a_ , embed=a_ ) , ) -> str: try: _UpperCAmelCase = self._pipeline.tokenizer.decode(a_ , a_ , a_ ) return ServeDeTokenizeResult(model="" , text=a_ ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(a_ )} ) async def _a ( self , a_=Body(a_ , embed=a_ ) ) -> Dict: # Check we don't have empty string if len(a_ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _UpperCAmelCase = self._pipeline(a_ ) return ServeForwardResult(output=a_ ) except Exception as e: raise HTTPException(500 , {"error": str(a_ )} )
657
"""simple docstring""" # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, nicht wahr?", } # BLUE scores as follows: # "pair": [fairseq, transformers] _UpperCAmelCase = { "wmt16-en-de-dist-12-1": [28.3, 27.52], "wmt16-en-de-dist-6-1": [27.4, 27.11], "wmt16-en-de-12-1": [26.9, 25.75], } _UpperCAmelCase = f"{src_lang}-{tgt_lang}" _UpperCAmelCase = f"\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"allenai/{model_name}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don't use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n" model_card_dir.mkdir(parents=UpperCamelCase__ , exist_ok=UpperCamelCase__ ) _UpperCAmelCase = os.path.join(UpperCamelCase__ , "README.md" ) print(f"Generating {path}" ) with open(UpperCamelCase__ , "w" , encoding="utf-8" ) as f: f.write(UpperCamelCase__ ) # make sure we are under the root of the project __magic_name__ = Path(__file__).resolve().parent.parent.parent __magic_name__ = repo_dir / '''model_cards''' for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: __magic_name__ = model_cards_dir / '''allenai''' / model_name write_model_card(model_card_dir, src_lang='''en''', tgt_lang='''de''', model_name=model_name)
657
1
"""simple docstring""" import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Optional[int] = (EulerDiscreteScheduler,) lowercase_ : List[Any] = 10 def _a ( self , **a_ ) -> int: _UpperCAmelCase = { "num_train_timesteps": 1100, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**a_ ) return config def _a ( self ) -> int: for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=a_ ) def _a ( self ) -> Optional[int]: for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=a_ , beta_end=a_ ) def _a ( self ) -> List[str]: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a_ ) def _a ( self ) -> Optional[int]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a_ ) def _a ( self ) -> Optional[Any]: _UpperCAmelCase = self.scheduler_classes[0] _UpperCAmelCase = self.get_scheduler_config() _UpperCAmelCase = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = self.dummy_model() _UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma _UpperCAmelCase = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase = scheduler.scale_model_input(a_ , a_ ) _UpperCAmelCase = model(a_ , a_ ) _UpperCAmelCase = scheduler.step(a_ , a_ , a_ , generator=a_ ) _UpperCAmelCase = output.prev_sample _UpperCAmelCase = torch.sum(torch.abs(a_ ) ) _UpperCAmelCase = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def _a ( self ) -> Dict: _UpperCAmelCase = self.scheduler_classes[0] _UpperCAmelCase = self.get_scheduler_config(prediction_type="v_prediction" ) _UpperCAmelCase = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = self.dummy_model() _UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma _UpperCAmelCase = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase = scheduler.scale_model_input(a_ , a_ ) _UpperCAmelCase = model(a_ , a_ ) _UpperCAmelCase = scheduler.step(a_ , a_ , a_ , generator=a_ ) _UpperCAmelCase = output.prev_sample _UpperCAmelCase = torch.sum(torch.abs(a_ ) ) _UpperCAmelCase = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 0.0002 ) < 1e-2 assert abs(result_mean.item() - 2.2_676e-06 ) < 1e-3 def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = self.scheduler_classes[0] _UpperCAmelCase = self.get_scheduler_config() _UpperCAmelCase = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps , device=a_ ) _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = self.dummy_model() _UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() _UpperCAmelCase = sample.to(a_ ) for t in scheduler.timesteps: _UpperCAmelCase = scheduler.scale_model_input(a_ , a_ ) _UpperCAmelCase = model(a_ , a_ ) _UpperCAmelCase = scheduler.step(a_ , a_ , a_ , generator=a_ ) _UpperCAmelCase = output.prev_sample _UpperCAmelCase = torch.sum(torch.abs(a_ ) ) _UpperCAmelCase = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def _a ( self ) -> List[str]: _UpperCAmelCase = self.scheduler_classes[0] _UpperCAmelCase = self.get_scheduler_config() _UpperCAmelCase = scheduler_class(**a_ , use_karras_sigmas=a_ ) scheduler.set_timesteps(self.num_inference_steps , device=a_ ) _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = self.dummy_model() _UpperCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() _UpperCAmelCase = sample.to(a_ ) for t in scheduler.timesteps: _UpperCAmelCase = scheduler.scale_model_input(a_ , a_ ) _UpperCAmelCase = model(a_ , a_ ) _UpperCAmelCase = scheduler.step(a_ , a_ , a_ , generator=a_ ) _UpperCAmelCase = output.prev_sample _UpperCAmelCase = torch.sum(torch.abs(a_ ) ) _UpperCAmelCase = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 124.52299499511719 ) < 1e-2 assert abs(result_mean.item() - 0.16213932633399963 ) < 1e-3
657
"""simple docstring""" from ..utils import DummyObject, requires_backends class _lowerCAmelCase ( metaclass=lowerCamelCase ): lowercase_ : Dict = ['''torch''', '''torchsde'''] def __init__( self , *a_ , **a_ ) -> Optional[int]: requires_backends(self , ["torch", "torchsde"] ) @classmethod def _a ( cls , *a_ , **a_ ) -> Optional[Any]: requires_backends(cls , ["torch", "torchsde"] ) @classmethod def _a ( cls , *a_ , **a_ ) -> List[Any]: requires_backends(cls , ["torch", "torchsde"] )
657
1