code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = '''▁''' UpperCAmelCase_ = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', '''tokenizer_config_file''': '''tokenizer_config.json''', } UpperCAmelCase_ = { '''vocab_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json''', }, '''spm_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model''', }, '''tokenizer_config_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json''', }, } UpperCAmelCase_ = { '''facebook/m2m100_418M''': 10_24, } # fmt: off UpperCAmelCase_ = { '''m2m100''': ['''af''', '''am''', '''ar''', '''ast''', '''az''', '''ba''', '''be''', '''bg''', '''bn''', '''br''', '''bs''', '''ca''', '''ceb''', '''cs''', '''cy''', '''da''', '''de''', '''el''', '''en''', '''es''', '''et''', '''fa''', '''ff''', '''fi''', '''fr''', '''fy''', '''ga''', '''gd''', '''gl''', '''gu''', '''ha''', '''he''', '''hi''', '''hr''', '''ht''', '''hu''', '''hy''', '''id''', '''ig''', '''ilo''', '''is''', '''it''', '''ja''', '''jv''', '''ka''', '''kk''', '''km''', '''kn''', '''ko''', '''lb''', '''lg''', '''ln''', '''lo''', '''lt''', '''lv''', '''mg''', '''mk''', '''ml''', '''mn''', '''mr''', '''ms''', '''my''', '''ne''', '''nl''', '''no''', '''ns''', '''oc''', '''or''', '''pa''', '''pl''', '''ps''', '''pt''', '''ro''', '''ru''', '''sd''', '''si''', '''sk''', '''sl''', '''so''', '''sq''', '''sr''', '''ss''', '''su''', '''sv''', '''sw''', '''ta''', '''th''', '''tl''', '''tn''', '''tr''', '''uk''', '''ur''', '''uz''', '''vi''', '''wo''', '''xh''', '''yi''', '''yo''', '''zh''', '''zu'''], '''wmt21''': ['''en''', '''ha''', '''is''', '''ja''', '''cs''', '''ru''', '''zh''', '''de'''] } class __UpperCamelCase ( A__ ): __A : Optional[Any] = VOCAB_FILES_NAMES __A : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __A : Any = ["input_ids", "attention_mask"] __A : List[int] = [] __A : List[int] = [] def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=None , _UpperCamelCase="<s>" , _UpperCamelCase="</s>" , _UpperCamelCase="</s>" , _UpperCamelCase="<pad>" , _UpperCamelCase="<unk>" , _UpperCamelCase="m2m100" , _UpperCamelCase = None , _UpperCamelCase=8 , **_UpperCamelCase , ): _UpperCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs _UpperCAmelCase = language_codes _UpperCAmelCase = FAIRSEQ_LANGUAGE_CODES[language_codes] _UpperCAmelCase = {lang_code: f'''__{lang_code}__''' for lang_code in fairseq_language_code} _UpperCAmelCase = kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(UpperCAmelCase__ ) for lang_code in fairseq_language_code if self.get_lang_token(UpperCAmelCase__ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=UpperCAmelCase__ , tgt_lang=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , language_codes=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=UpperCAmelCase__ , **UpperCAmelCase__ , ) _UpperCAmelCase = vocab_file _UpperCAmelCase = load_json(UpperCAmelCase__ ) _UpperCAmelCase = {v: k for k, v in self.encoder.items()} _UpperCAmelCase = spm_file _UpperCAmelCase = load_spm(UpperCAmelCase__ , self.sp_model_kwargs ) _UpperCAmelCase = len(self.encoder ) _UpperCAmelCase = { self.get_lang_token(UpperCAmelCase__ ): self.encoder_size + i for i, lang_code in enumerate(UpperCAmelCase__ ) } _UpperCAmelCase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(UpperCAmelCase__ )} _UpperCAmelCase = {v: k for k, v in self.lang_token_to_id.items()} _UpperCAmelCase = src_lang if src_lang is not None else '''en''' _UpperCAmelCase = tgt_lang _UpperCAmelCase = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) _UpperCAmelCase = num_madeup_words @property def UpperCamelCase( self ): return len(self.encoder ) + len(self.lang_token_to_id ) @property def UpperCamelCase( self ): return self._src_lang @src_lang.setter def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase( self , _UpperCamelCase ): return self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) def UpperCamelCase( self , _UpperCamelCase ): if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(UpperCAmelCase__ , self.encoder[self.unk_token] ) def UpperCamelCase( self , _UpperCamelCase ): if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(UpperCAmelCase__ , self.unk_token ) def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = [] _UpperCAmelCase = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCAmelCase__ ) + token _UpperCAmelCase = [] else: current_sub_tokens.append(UpperCAmelCase__ ) out_string += self.sp_model.decode(UpperCAmelCase__ ) return out_string.strip() def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) _UpperCAmelCase = [1] * len(self.prefix_tokens ) _UpperCAmelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(UpperCAmelCase__ )) + suffix_ones return prefix_ones + ([0] * len(UpperCAmelCase__ )) + ([0] * len(UpperCAmelCase__ )) + suffix_ones def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase( self ): _UpperCAmelCase = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): _UpperCAmelCase = self.__dict__.copy() _UpperCAmelCase = None return state def __setstate__( self , _UpperCamelCase ): _UpperCAmelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _UpperCAmelCase = {} _UpperCAmelCase = load_spm(self.spm_file , self.sp_model_kwargs ) def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase = None ): _UpperCAmelCase = Path(UpperCAmelCase__ ) if not save_dir.is_dir(): raise OSError(f'''{save_directory} should be a directory''' ) _UpperCAmelCase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) _UpperCAmelCase = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , UpperCAmelCase__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , UpperCAmelCase__ ) elif not os.path.isfile(self.spm_file ): with open(UpperCAmelCase__ , '''wb''' ) as fi: _UpperCAmelCase = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (str(UpperCAmelCase__ ), str(UpperCAmelCase__ )) def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase = "en" , _UpperCamelCase = None , _UpperCamelCase = "ro" , **_UpperCamelCase , ): _UpperCAmelCase = src_lang _UpperCAmelCase = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCamelCase( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ): if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) _UpperCAmelCase = src_lang _UpperCAmelCase = self(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ ) _UpperCAmelCase = self.get_lang_id(UpperCAmelCase__ ) _UpperCAmelCase = tgt_lang_id return inputs def UpperCamelCase( self ): self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase( self ): self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = self.get_lang_token(UpperCAmelCase__ ) _UpperCAmelCase = self.lang_token_to_id[lang_token] _UpperCAmelCase = [self.cur_lang_id] _UpperCAmelCase = [self.eos_token_id] def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = self.get_lang_token(UpperCAmelCase__ ) _UpperCAmelCase = self.lang_token_to_id[lang_token] _UpperCAmelCase = [self.cur_lang_id] _UpperCAmelCase = [self.eos_token_id] def UpperCamelCase( self , _UpperCamelCase ): return self.lang_code_to_token[lang] def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = self.get_lang_token(UpperCAmelCase__ ) return self.lang_token_to_id[lang_token] def A__ ( SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Tuple ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = sentencepiece.SentencePieceProcessor(**lowerCAmelCase_ ) spm.Load(str(lowerCAmelCase_ ) ) return spm def A__ ( SCREAMING_SNAKE_CASE_ : Dict ) -> Tuple: """simple docstring""" with open(lowerCAmelCase_ , '''r''' ) as f: return json.load(lowerCAmelCase_ ) def A__ ( SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Optional[Any]: """simple docstring""" with open(lowerCAmelCase_ , '''w''' ) as f: json.dump(lowerCAmelCase_ , lowerCAmelCase_ , indent=2 )
32
"""simple docstring""" import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : """simple docstring""" def __init__( self : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=1_3 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Tuple=9_9 , UpperCAmelCase__ : Dict=3_2 , UpperCAmelCase__ : int=5 , UpperCAmelCase__ : str=4 , UpperCAmelCase__ : List[Any]=3_7 , UpperCAmelCase__ : int="gelu" , UpperCAmelCase__ : List[Any]=0.1 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Tuple=1_2_8 , UpperCAmelCase__ : Union[str, Any]=3_2 , UpperCAmelCase__ : Any=1_6 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Tuple=0.02 , UpperCAmelCase__ : str=3 , UpperCAmelCase__ : Tuple=4 , UpperCAmelCase__ : List[str]=None , ) -> Optional[int]: __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = scope def UpperCAmelCase_ ( self : str ) -> Any: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_input_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: ( ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ) = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Union[str, Any] ) -> Any: __SCREAMING_SNAKE_CASE = NezhaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , ) -> Tuple: __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = NezhaModel(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , encoder_attention_mask=UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] ) -> int: __SCREAMING_SNAKE_CASE = NezhaForMaskedLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any ) -> Tuple: __SCREAMING_SNAKE_CASE = NezhaForNextSentencePrediction(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] ) -> List[str]: __SCREAMING_SNAKE_CASE = NezhaForPreTraining(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , next_sentence_label=UpperCAmelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = NezhaForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Any ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = NezhaForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] ) -> Any: __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = NezhaForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Dict ) -> str: __SCREAMING_SNAKE_CASE = self.num_choices __SCREAMING_SNAKE_CASE = NezhaForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __SCREAMING_SNAKE_CASE = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __SCREAMING_SNAKE_CASE = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self : List[str] ) -> List[str]: __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ) = config_and_inputs __SCREAMING_SNAKE_CASE = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : str = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) snake_case__ : Tuple = ( { "feature-extraction": NezhaModel, "fill-mask": NezhaForMaskedLM, "question-answering": NezhaForQuestionAnswering, "text-classification": NezhaForSequenceClassification, "token-classification": NezhaForTokenClassification, "zero-shot": NezhaForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : int = True def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any]=False ) -> Dict: __SCREAMING_SNAKE_CASE = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) if return_labels: if model_class in get_values(UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase__ ) return inputs_dict def UpperCAmelCase_ ( self : List[str] ) -> List[str]: __SCREAMING_SNAKE_CASE = NezhaModelTester(self ) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=3_7 ) def UpperCAmelCase_ ( self : int ) -> List[Any]: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[str] ) -> Dict: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ) -> Dict: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> List[Any]: # This regression test was failing with PyTorch < 1.3 ( ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() __SCREAMING_SNAKE_CASE = None self.model_tester.create_and_check_model_as_decoder( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase__ ) @slow def UpperCAmelCase_ ( self : List[Any] ) -> int: for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = NezhaModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @slow @require_torch_gpu def UpperCAmelCase_ ( self : List[str] ) -> Optional[int]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = model_class(config=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.jit.trace( UpperCAmelCase__ , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(UpperCAmelCase__ , os.path.join(UpperCAmelCase__ , "bert.pt" ) ) __SCREAMING_SNAKE_CASE = torch.jit.load(os.path.join(UpperCAmelCase__ , "bert.pt" ) , map_location=UpperCAmelCase__ ) loaded(inputs_dict["input_ids"].to(UpperCAmelCase__ ) , inputs_dict["attention_mask"].to(UpperCAmelCase__ ) ) @require_torch class UpperCamelCase_ ( unittest.TestCase): """simple docstring""" @slow def UpperCAmelCase_ ( self : List[Any] ) -> str: __SCREAMING_SNAKE_CASE = NezhaModel.from_pretrained("sijunhe/nezha-cn-base" ) __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 6, 7_6_8) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.tensor([[[0.0_685, 0.2_441, 0.1_102], [0.0_600, 0.1_906, 0.1_349], [0.0_221, 0.0_819, 0.0_586]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCAmelCase__ , atol=1E-4 ) ) @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: __SCREAMING_SNAKE_CASE = NezhaForMaskedLM.from_pretrained("sijunhe/nezha-cn-base" ) __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __SCREAMING_SNAKE_CASE = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 6, 2_1_1_2_8) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.tensor( [[-2.7_939, -1.7_902, -2.2_189], [-2.8_585, -1.8_908, -2.3_723], [-2.6_499, -1.7_750, -2.2_558]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCAmelCase__ , atol=1E-4 ) )
682
0
import re def a ( lowerCamelCase_ ): '''simple docstring''' lowercase__ = re.compile(r'''^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$''' ) if match := re.search(lowerCAmelCase_ , lowerCAmelCase_ ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
183
"""simple docstring""" import os def UpperCAmelCase__ (): '''simple docstring''' with open(os.path.dirname(lowerCAmelCase_ ) + "/p022_names.txt" ) as file: __SCREAMING_SNAKE_CASE = str(file.readlines()[0] ) __SCREAMING_SNAKE_CASE = names.replace("\"" , "" ).split("," ) names.sort() __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 for i, name in enumerate(lowerCAmelCase_ ): for letter in name: name_score += ord(lowerCAmelCase_ ) - 64 total_score += (i + 1) * name_score __SCREAMING_SNAKE_CASE = 0 return total_score if __name__ == "__main__": print(solution())
682
0
'''simple docstring''' import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __A ( UpperCamelCase__ ): def __init__(self : str , __a : Tuple , __a : str , __a : Tuple=1024 , __a : Dict=1024 , __a : int=3.6 ): UpperCAmelCase_ = tokenizer UpperCAmelCase_ = tokenizer.bos_token_id UpperCAmelCase_ = dataset UpperCAmelCase_ = seq_length UpperCAmelCase_ = seq_length * chars_per_token * num_of_sequences def __iter__(self : str ): UpperCAmelCase_ = iter(self.dataset ) UpperCAmelCase_ = True while more_examples: UpperCAmelCase_ , UpperCAmelCase_ = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(UpperCAmelCase__ )["content"] ) buffer_len += len(buffer[-1] ) except StopIteration: UpperCAmelCase_ = False break UpperCAmelCase_ = tokenizer(UpperCAmelCase__ , truncation=UpperCAmelCase__ )["input_ids"] UpperCAmelCase_ = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(UpperCAmelCase__ ) , self.seq_length ): UpperCAmelCase_ = all_token_ids[i : i + self.seq_length] if len(UpperCAmelCase__ ) == self.seq_length: yield torch.tensor(UpperCAmelCase__ ) def lowerCAmelCase_ ( snake_case_ : str ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = {"streaming": True} UpperCAmelCase_ = load_dataset(args.dataset_name , split="train" , **lowerCAmelCase_ ) UpperCAmelCase_ = ConstantLengthDataset(lowerCAmelCase_ , lowerCAmelCase_ , seq_length=args.seq_length ) UpperCAmelCase_ = DataLoader(lowerCAmelCase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_ ( snake_case_ : Dict ) -> str: '''simple docstring''' model.eval() UpperCAmelCase_ = [] for step, batch in enumerate(lowerCAmelCase_ ): with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase_ , labels=lowerCAmelCase_ ) UpperCAmelCase_ = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowerCAmelCase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break UpperCAmelCase_ = torch.mean(torch.cat(lowerCAmelCase_ ) ) try: UpperCAmelCase_ = torch.exp(lowerCAmelCase_ ) except OverflowError: UpperCAmelCase_ = float("inf" ) return loss.item(), perplexity.item() # Setup Accelerator SCREAMING_SNAKE_CASE_: Tuple =Accelerator() # Parse configuration SCREAMING_SNAKE_CASE_: Tuple =HfArgumentParser(EvaluationArguments) SCREAMING_SNAKE_CASE_: Optional[Any] =parser.parse_args() set_seed(args.seed) # Logging SCREAMING_SNAKE_CASE_: Optional[int] =logging.getLogger(__name__) logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) # Load model and tokenizer SCREAMING_SNAKE_CASE_: Optional[Any] =AutoModelForCausalLM.from_pretrained(args.model_ckpt) SCREAMING_SNAKE_CASE_: int =AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader SCREAMING_SNAKE_CASE_: List[Any] =create_dataloader(args) # Prepare everything with our `accelerator`. SCREAMING_SNAKE_CASE_: Tuple =accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('Evaluating and saving model after training') SCREAMING_SNAKE_CASE_: Union[str, Any] =evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
78
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 1.5 __SCREAMING_SNAKE_CASE = int(factor * num_class_images ) __SCREAMING_SNAKE_CASE = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=lowerCAmelCase_ , aesthetic_weight=0.1 ) os.makedirs(f"""{class_data_dir}/images""" , exist_ok=lowerCAmelCase_ ) if len(list(Path(f"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: __SCREAMING_SNAKE_CASE = client.query(text=lowerCAmelCase_ ) if len(lowerCAmelCase_ ) >= factor * num_class_images or num_images > 1E4: break else: __SCREAMING_SNAKE_CASE = int(factor * num_images ) __SCREAMING_SNAKE_CASE = ClipClient( url="https://knn.laion.ai/knn-service" , indice_name="laion_400m" , num_images=lowerCAmelCase_ , aesthetic_weight=0.1 , ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = tqdm(desc="downloading real regularization images" , total=lowerCAmelCase_ ) with open(f"""{class_data_dir}/caption.txt""" , "w" ) as fa, open(f"""{class_data_dir}/urls.txt""" , "w" ) as fa, open( f"""{class_data_dir}/images.txt""" , "w" ) as fa: while total < num_class_images: __SCREAMING_SNAKE_CASE = class_images[count] count += 1 try: __SCREAMING_SNAKE_CASE = requests.get(images["url"] ) if img.status_code == 200: __SCREAMING_SNAKE_CASE = Image.open(BytesIO(img.content ) ) with open(f"""{class_data_dir}/images/{total}.jpg""" , "wb" ) as f: f.write(img.content ) fa.write(images["caption"] + "\n" ) fa.write(images["url"] + "\n" ) fa.write(f"""{class_data_dir}/images/{total}.jpg""" + "\n" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def UpperCAmelCase__ (): '''simple docstring''' __SCREAMING_SNAKE_CASE = argparse.ArgumentParser("" , add_help=lowerCAmelCase_ ) parser.add_argument("--class_prompt" , help="text prompt to retrieve images" , required=lowerCAmelCase_ , type=lowerCAmelCase_ ) parser.add_argument("--class_data_dir" , help="path to save images" , required=lowerCAmelCase_ , type=lowerCAmelCase_ ) parser.add_argument("--num_class_images" , help="number of images to download" , default=200 , type=lowerCAmelCase_ ) return parser.parse_args() if __name__ == "__main__": a__ : Optional[Any] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
682
0
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def UpperCamelCase ( snake_case__ : Dict ,snake_case__ : Tuple ): '''simple docstring''' assert isinstance(lowerCAmelCase_ ,lowerCAmelCase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" ,[False, True] ) def UpperCamelCase ( snake_case__ : Dict ,snake_case__ : int ,snake_case__ : str ,snake_case__ : List[Any] ): '''simple docstring''' __snake_case :Optional[int] = tmp_path / """cache""" __snake_case :int = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __snake_case :str = SqlDatasetReader( """dataset""" ,"""sqlite:///""" + sqlite_path ,cache_dir=lowerCAmelCase_ ,keep_in_memory=lowerCAmelCase_ ).read() _check_sql_dataset(lowerCAmelCase_ ,lowerCAmelCase_ ) @require_sqlalchemy @pytest.mark.parametrize( """features""" ,[ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] ,) def UpperCamelCase ( snake_case__ : int ,snake_case__ : List[Any] ,snake_case__ : List[Any] ,snake_case__ : Optional[Any] ): '''simple docstring''' __snake_case :List[Any] = tmp_path / """cache""" __snake_case :Optional[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __snake_case :Optional[Any] = features.copy() if features else default_expected_features __snake_case :Dict = ( Features({feature: Value(lowerCAmelCase_ ) for feature, dtype in features.items()} ) if features is not None else None ) __snake_case :Optional[Any] = SqlDatasetReader("""dataset""" ,"""sqlite:///""" + sqlite_path ,features=lowerCAmelCase_ ,cache_dir=lowerCAmelCase_ ).read() _check_sql_dataset(lowerCAmelCase_ ,lowerCAmelCase_ ) def UpperCamelCase ( snake_case__ : str ): '''simple docstring''' with contextlib.closing(sqlitea.connect(lowerCAmelCase_ ) ) as con: __snake_case :Union[str, Any] = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def UpperCamelCase ( snake_case__ : str ,snake_case__ : List[Any] ,snake_case__ : Dict ): '''simple docstring''' __snake_case :Dict = tmp_path / """cache""" __snake_case :Optional[int] = os.path.join(lowerCAmelCase_ ,"""tmp.sql""" ) __snake_case :Any = SqlDatasetReader("""dataset""" ,"""sqlite:///""" + sqlite_path ,cache_dir=lowerCAmelCase_ ).read() SqlDatasetWriter(lowerCAmelCase_ ,"""dataset""" ,"""sqlite:///""" + output_sqlite_path ,num_proc=1 ).write() __snake_case :Dict = iter_sql_file(lowerCAmelCase_ ) __snake_case :int = iter_sql_file(lowerCAmelCase_ ) for rowa, rowa in zip(lowerCAmelCase_ ,lowerCAmelCase_ ): assert rowa == rowa @require_sqlalchemy def UpperCamelCase ( snake_case__ : List[str] ,snake_case__ : Union[str, Any] ,snake_case__ : Union[str, Any] ): '''simple docstring''' __snake_case :List[Any] = tmp_path / """cache""" __snake_case :List[str] = os.path.join(lowerCAmelCase_ ,"""tmp.sql""" ) __snake_case :Optional[int] = SqlDatasetReader("""dataset""" ,"""sqlite:///""" + sqlite_path ,cache_dir=lowerCAmelCase_ ).read() SqlDatasetWriter(lowerCAmelCase_ ,"""dataset""" ,"""sqlite:///""" + output_sqlite_path ,num_proc=2 ).write() __snake_case :Union[str, Any] = iter_sql_file(lowerCAmelCase_ ) __snake_case :List[Any] = iter_sql_file(lowerCAmelCase_ ) for rowa, rowa in zip(lowerCAmelCase_ ,lowerCAmelCase_ ): assert rowa == rowa @require_sqlalchemy def UpperCamelCase ( snake_case__ : Union[str, Any] ,snake_case__ : List[str] ,snake_case__ : Dict ): '''simple docstring''' __snake_case :int = tmp_path / """cache""" __snake_case :List[Any] = os.path.join(lowerCAmelCase_ ,"""tmp.sql""" ) __snake_case :int = SqlDatasetReader("""dataset""" ,"""sqlite:///""" + sqlite_path ,cache_dir=lowerCAmelCase_ ).read() with pytest.raises(lowerCAmelCase_ ): SqlDatasetWriter(lowerCAmelCase_ ,"""dataset""" ,"""sqlite:///""" + output_sqlite_path ,num_proc=0 ).write()
455
"""simple docstring""" import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING a__ : str = logging.get_logger(__name__) class UpperCamelCase_ ( enum.Enum): """simple docstring""" snake_case__ : Optional[int] = 0 snake_case__ : Dict = 1 @add_end_docstrings(UpperCamelCase) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Tuple = "generated" def __init__( self : Any , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : str ) -> Dict: super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Dict=None , **UpperCAmelCase__ : Union[str, Any] , ) -> Optional[int]: __SCREAMING_SNAKE_CASE = {} if truncation is not None: __SCREAMING_SNAKE_CASE = truncation __SCREAMING_SNAKE_CASE = generate_kwargs __SCREAMING_SNAKE_CASE = {} if return_tensors is not None and return_type is None: __SCREAMING_SNAKE_CASE = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: __SCREAMING_SNAKE_CASE = return_type if clean_up_tokenization_spaces is not None: __SCREAMING_SNAKE_CASE = clean_up_tokenization_spaces if stop_sequence is not None: __SCREAMING_SNAKE_CASE = self.tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) __SCREAMING_SNAKE_CASE = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> List[str]: return True def UpperCAmelCase_ ( self : Any , *UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] ) -> Any: __SCREAMING_SNAKE_CASE = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , UpperCAmelCase__ ): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input" ) __SCREAMING_SNAKE_CASE = ([prefix + arg for arg in args[0]],) __SCREAMING_SNAKE_CASE = True elif isinstance(args[0] , UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = (prefix + args[0],) __SCREAMING_SNAKE_CASE = False else: raise ValueError( F""" `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`""" ) __SCREAMING_SNAKE_CASE = self.tokenizer(*UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : List[str] , *UpperCAmelCase__ : Any , **UpperCAmelCase__ : Union[str, Any] ) -> List[Any]: __SCREAMING_SNAKE_CASE = super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ ) if ( isinstance(args[0] , UpperCAmelCase__ ) and all(isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) for el in args[0] ) and all(len(UpperCAmelCase__ ) == 1 for res in result ) ): return [res[0] for res in result] return result def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int]=TruncationStrategy.DO_NOT_TRUNCATE , **UpperCAmelCase__ : int ) -> Tuple: __SCREAMING_SNAKE_CASE = self._parse_and_tokenize(UpperCAmelCase__ , truncation=UpperCAmelCase__ , **UpperCAmelCase__ ) return inputs def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : int , **UpperCAmelCase__ : Any ) -> Any: if self.framework == "pt": __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = model_inputs["input_ids"].shape elif self.framework == "tf": __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = tf.shape(model_inputs["input_ids"] ).numpy() __SCREAMING_SNAKE_CASE = generate_kwargs.get("min_length" , self.model.config.min_length ) __SCREAMING_SNAKE_CASE = generate_kwargs.get("max_length" , self.model.config.max_length ) self.check_inputs(UpperCAmelCase__ , generate_kwargs["min_length"] , generate_kwargs["max_length"] ) __SCREAMING_SNAKE_CASE = self.model.generate(**UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = output_ids.shape[0] if self.framework == "pt": __SCREAMING_SNAKE_CASE = output_ids.reshape(UpperCAmelCase__ , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": __SCREAMING_SNAKE_CASE = tf.reshape(UpperCAmelCase__ , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict=ReturnType.TEXT , UpperCAmelCase__ : str=False ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: __SCREAMING_SNAKE_CASE = {F"""{self.return_name}_token_ids""": output_ids} elif return_type == ReturnType.TEXT: __SCREAMING_SNAKE_CASE = { F"""{self.return_name}_text""": self.tokenizer.decode( UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ , clean_up_tokenization_spaces=UpperCAmelCase__ , ) } records.append(UpperCAmelCase__ ) return records @add_end_docstrings(UpperCamelCase) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : str = "summary" def __call__( self : Tuple , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Any ) -> Optional[int]: return super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> bool: if max_length < min_length: logger.warning(F"""Your min_length={min_length} must be inferior than your max_length={max_length}.""" ) if input_length < max_length: logger.warning( F"""Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is """ "a summarization task, where outputs shorter than the input are typically wanted, you might " F"""consider decreasing max_length manually, e.g. summarizer('...', max_length={input_length//2})""" ) @add_end_docstrings(UpperCamelCase) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : str = "translation" def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> Optional[Any]: if input_length > 0.9 * max_length: logger.warning( F"""Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider """ "increasing your max_length manually, e.g. translator('...', max_length=400)" ) return True def UpperCAmelCase_ ( self : Any , *UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int]=TruncationStrategy.DO_NOT_TRUNCATE , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Optional[Any]=None ) -> List[Any]: if getattr(self.tokenizer , "_build_translation_inputs" , UpperCAmelCase__ ): return self.tokenizer._build_translation_inputs( *UpperCAmelCase__ , return_tensors=self.framework , truncation=UpperCAmelCase__ , src_lang=UpperCAmelCase__ , tgt_lang=UpperCAmelCase__ ) else: return super()._parse_and_tokenize(*UpperCAmelCase__ , truncation=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : str=None , **UpperCAmelCase__ : List[str] ) -> Any: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = super()._sanitize_parameters(**UpperCAmelCase__ ) if src_lang is not None: __SCREAMING_SNAKE_CASE = src_lang if tgt_lang is not None: __SCREAMING_SNAKE_CASE = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. __SCREAMING_SNAKE_CASE = kwargs.get("task" , self.task ) __SCREAMING_SNAKE_CASE = task.split("_" ) if task and len(UpperCAmelCase__ ) == 4: # translation, XX, to YY __SCREAMING_SNAKE_CASE = items[1] __SCREAMING_SNAKE_CASE = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : str , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Any ) -> List[Any]: return super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ )
682
0
from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline _lowerCAmelCase: Union[str, Any] = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowercase_ (lowercase__ ): def __init__( self , **lowercase_) -> Dict: super().__init__(**UpperCAmelCase__) if self.framework != "pt": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""") # No specific FOR_XXX available yet def __call__( self , lowercase_ , **lowercase_) -> List[Any]: return super().__call__(UpperCAmelCase__ , **UpperCAmelCase__) def __UpperCamelCase ( self , **lowercase_) -> Any: a__ ={} if "candidate_labels" in kwargs: a__ =kwargs['candidate_labels'] if "hypothesis_template" in kwargs: a__ =kwargs['hypothesis_template'] return preprocess_params, {}, {} def __UpperCamelCase ( self , lowercase_ , lowercase_=None , lowercase_="This is a sound of {}.") -> List[Any]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__): if audio.startswith('http://') or audio.startswith('https://'): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png a__ =requests.get(UpperCAmelCase__).content else: with open(UpperCAmelCase__ , 'rb') as f: a__ =f.read() if isinstance(UpperCAmelCase__ , UpperCAmelCase__): a__ =ffmpeg_read(UpperCAmelCase__ , self.feature_extractor.sampling_rate) if not isinstance(UpperCAmelCase__ , np.ndarray): raise ValueError('We expect a numpy ndarray as input') if len(audio.shape) != 1: raise ValueError('We expect a single channel audio input for ZeroShotAudioClassificationPipeline') a__ =self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors='pt') a__ =candidate_labels a__ =[hypothesis_template.format(UpperCAmelCase__) for x in candidate_labels] a__ =self.tokenizer(UpperCAmelCase__ , return_tensors=self.framework , padding=UpperCAmelCase__) a__ =[text_inputs] return inputs def __UpperCamelCase ( self , lowercase_) -> str: a__ =model_inputs.pop('candidate_labels') a__ =model_inputs.pop('text_inputs') if isinstance(text_inputs[0] , UpperCAmelCase__): a__ =text_inputs[0] else: # Batching case. a__ =text_inputs[0][0] a__ =self.model(**UpperCAmelCase__ , **UpperCAmelCase__) a__ ={ 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_audio, } return model_outputs def __UpperCamelCase ( self , lowercase_) -> Union[str, Any]: a__ =model_outputs.pop('candidate_labels') a__ =model_outputs['logits'][0] if self.framework == "pt": a__ =logits.softmax(dim=0) a__ =probs.tolist() else: raise ValueError('`tf` framework not supported.') a__ =[ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(UpperCAmelCase__ , UpperCAmelCase__) , key=lambda lowercase_: -x[0]) ] return result
20
"""simple docstring""" import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : List[Any] = AutoencoderKL snake_case__ : Optional[Any] = "sample" snake_case__ : Optional[Any] = 1E-2 @property def UpperCAmelCase_ ( self : Tuple ) -> int: __SCREAMING_SNAKE_CASE = 4 __SCREAMING_SNAKE_CASE = 3 __SCREAMING_SNAKE_CASE = (3_2, 3_2) __SCREAMING_SNAKE_CASE = floats_tensor((batch_size, num_channels) + sizes ).to(UpperCAmelCase__ ) return {"sample": image} @property def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[int]: return (3, 3_2, 3_2) @property def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: return (3, 3_2, 3_2) def UpperCAmelCase_ ( self : List[str] ) -> List[str]: __SCREAMING_SNAKE_CASE = { "block_out_channels": [3_2, 6_4], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __SCREAMING_SNAKE_CASE = self.dummy_input return init_dict, inputs_dict def UpperCAmelCase_ ( self : Optional[Any] ) -> List[str]: pass def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def UpperCAmelCase_ ( self : str ) -> List[Any]: # enable deterministic behavior for gradient checkpointing __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.prepare_init_args_and_inputs_for_common() __SCREAMING_SNAKE_CASE = self.model_class(**UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) assert not model.is_gradient_checkpointing and model.training __SCREAMING_SNAKE_CASE = model(**UpperCAmelCase__ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __SCREAMING_SNAKE_CASE = torch.randn_like(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __SCREAMING_SNAKE_CASE = self.model_class(**UpperCAmelCase__ ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(UpperCAmelCase__ ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __SCREAMING_SNAKE_CASE = model_a(**UpperCAmelCase__ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __SCREAMING_SNAKE_CASE = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __SCREAMING_SNAKE_CASE = dict(model.named_parameters() ) __SCREAMING_SNAKE_CASE = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def UpperCAmelCase_ ( self : List[str] ) -> Any: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __SCREAMING_SNAKE_CASE = model.to(UpperCAmelCase__ ) model.eval() if torch_device == "mps": __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) else: __SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) __SCREAMING_SNAKE_CASE = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __SCREAMING_SNAKE_CASE = image.to(UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , sample_posterior=UpperCAmelCase__ , generator=UpperCAmelCase__ ).sample __SCREAMING_SNAKE_CASE = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __SCREAMING_SNAKE_CASE = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __SCREAMING_SNAKE_CASE = torch.tensor( [-0.1_352, 0.0_878, 0.0_419, -0.0_818, -0.1_069, 0.0_688, -0.1_458, -0.4_446, -0.0_026] ) else: __SCREAMING_SNAKE_CASE = torch.tensor( [-0.2_421, 0.4_642, 0.2_507, -0.0_438, 0.0_682, 0.3_160, -0.2_018, -0.0_727, 0.2_485] ) self.assertTrue(torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , rtol=1E-2 ) ) @slow class UpperCamelCase_ ( unittest.TestCase): """simple docstring""" def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ) -> Any: return F"""gaussian_noise_s={seed}_shape={'_'.join([str(UpperCAmelCase__ ) for s in shape] )}.npy""" def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : Tuple=0 , UpperCAmelCase__ : Optional[Any]=(4, 3, 5_1_2, 5_1_2) , UpperCAmelCase__ : Any=False ) -> List[str]: __SCREAMING_SNAKE_CASE = torch.floataa if fpaa else torch.floataa __SCREAMING_SNAKE_CASE = torch.from_numpy(load_hf_numpy(self.get_file_format(UpperCAmelCase__ , UpperCAmelCase__ ) ) ).to(UpperCAmelCase__ ).to(UpperCAmelCase__ ) return image def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Dict="CompVis/stable-diffusion-v1-4" , UpperCAmelCase__ : Optional[Any]=False ) -> Tuple: __SCREAMING_SNAKE_CASE = "fp16" if fpaa else None __SCREAMING_SNAKE_CASE = torch.floataa if fpaa else torch.floataa __SCREAMING_SNAKE_CASE = AutoencoderKL.from_pretrained( UpperCAmelCase__ , subfolder="vae" , torch_dtype=UpperCAmelCase__ , revision=UpperCAmelCase__ , ) model.to(UpperCAmelCase__ ).eval() return model def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : int=0 ) -> str: if torch_device == "mps": return torch.manual_seed(UpperCAmelCase__ ) return torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) @parameterized.expand( [ # fmt: off [3_3, [-0.1_603, 0.9_878, -0.0_495, -0.0_790, -0.2_709, 0.8_375, -0.2_060, -0.0_824], [-0.2_395, 0.0_098, 0.0_102, -0.0_709, -0.2_840, -0.0_274, -0.0_718, -0.1_824]], [4_7, [-0.2_376, 0.1_168, 0.1_332, -0.4_840, -0.2_508, -0.0_791, -0.0_493, -0.4_089], [0.0_350, 0.0_847, 0.0_467, 0.0_344, -0.0_842, -0.0_547, -0.0_633, -0.1_131]], # fmt: on ] ) def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[str] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model() __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_generator(UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , generator=UpperCAmelCase__ , sample_posterior=UpperCAmelCase__ ).sample assert sample.shape == image.shape __SCREAMING_SNAKE_CASE = sample[-1, -2:, -2:, :2].flatten().float().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=3E-3 ) @parameterized.expand( [ # fmt: off [3_3, [-0.0_513, 0.0_289, 1.3_799, 0.2_166, -0.2_573, -0.0_871, 0.5_103, -0.0_999]], [4_7, [-0.4_128, -0.1_320, -0.3_704, 0.1_965, -0.4_116, -0.2_332, -0.3_340, 0.2_247]], # fmt: on ] ) @require_torch_gpu def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : int ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model(fpaa=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ , fpaa=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_generator(UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , generator=UpperCAmelCase__ , sample_posterior=UpperCAmelCase__ ).sample assert sample.shape == image.shape __SCREAMING_SNAKE_CASE = sample[-1, -2:, :2, -2:].flatten().float().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(UpperCAmelCase__ ) assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.1_609, 0.9_866, -0.0_487, -0.0_777, -0.2_716, 0.8_368, -0.2_055, -0.0_814], [-0.2_395, 0.0_098, 0.0_102, -0.0_709, -0.2_840, -0.0_274, -0.0_718, -0.1_824]], [4_7, [-0.2_377, 0.1_147, 0.1_333, -0.4_841, -0.2_506, -0.0_805, -0.0_491, -0.4_085], [0.0_350, 0.0_847, 0.0_467, 0.0_344, -0.0_842, -0.0_547, -0.0_633, -0.1_131]], # fmt: on ] ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any ) -> Dict: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model() __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ ).sample assert sample.shape == image.shape __SCREAMING_SNAKE_CASE = sample[-1, -2:, -2:, :2].flatten().float().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=3E-3 ) @parameterized.expand( [ # fmt: off [1_3, [-0.2_051, -0.1_803, -0.2_311, -0.2_114, -0.3_292, -0.3_574, -0.2_953, -0.3_323]], [3_7, [-0.2_632, -0.2_625, -0.2_199, -0.2_741, -0.4_539, -0.4_990, -0.3_720, -0.4_925]], # fmt: on ] ) @require_torch_gpu def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int ) -> str: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model() __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] __SCREAMING_SNAKE_CASE = sample[-1, -2:, :2, -2:].flatten().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(UpperCAmelCase__ ) assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-3 ) @parameterized.expand( [ # fmt: off [2_7, [-0.0_369, 0.0_207, -0.0_776, -0.0_682, -0.1_747, -0.1_930, -0.1_465, -0.2_039]], [1_6, [-0.1_628, -0.2_134, -0.2_747, -0.2_642, -0.3_774, -0.4_404, -0.3_687, -0.4_277]], # fmt: on ] ) @require_torch_gpu def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any ) -> Dict: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model(fpaa=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ , shape=(3, 4, 6_4, 6_4) , fpaa=UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] __SCREAMING_SNAKE_CASE = sample[-1, -2:, :2, -2:].flatten().float().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(UpperCAmelCase__ ) assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=5E-3 ) @parameterized.expand([(1_3,), (1_6,), (2_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : Union[str, Any] ) -> List[str]: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model(fpaa=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ , shape=(3, 4, 6_4, 6_4) , fpaa=UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-1 ) @parameterized.expand([(1_3,), (1_6,), (3_7,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : Tuple ) -> Dict: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model() __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ , shape=(3, 4, 6_4, 6_4) ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.decode(UpperCAmelCase__ ).sample assert list(sample.shape ) == [3, 3, 5_1_2, 5_1_2] assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [3_3, [-0.3_001, 0.0_918, -2.6_984, -3.9_720, -3.2_099, -5.0_353, 1.7_338, -0.2_065, 3.4_267]], [4_7, [-1.5_030, -4.3_871, -6.0_355, -9.1_157, -1.6_661, -2.7_853, 2.1_607, -5.0_823, 2.5_633]], # fmt: on ] ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Tuple ) -> Optional[int]: __SCREAMING_SNAKE_CASE = self.get_sd_vae_model() __SCREAMING_SNAKE_CASE = self.get_sd_image(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.get_generator(UpperCAmelCase__ ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model.encode(UpperCAmelCase__ ).latent_dist __SCREAMING_SNAKE_CASE = dist.sample(generator=UpperCAmelCase__ ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __SCREAMING_SNAKE_CASE = sample[0, -1, -3:, -3:].flatten().cpu() __SCREAMING_SNAKE_CASE = torch.tensor(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(UpperCAmelCase__ , UpperCAmelCase__ , atol=UpperCAmelCase__ )
682
0
from ... import PretrainedConfig lowercase_ = { '''sijunhe/nezha-cn-base''': '''https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json''', } class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' _A = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP _A = "nezha" def __init__( self : List[str] , SCREAMING_SNAKE_CASE_ : List[Any]=2_1_1_2_8 , SCREAMING_SNAKE_CASE_ : str=7_6_8 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=1_2 , SCREAMING_SNAKE_CASE_ : Optional[Any]=1_2 , SCREAMING_SNAKE_CASE_ : int=3_0_7_2 , SCREAMING_SNAKE_CASE_ : Union[str, Any]="gelu" , SCREAMING_SNAKE_CASE_ : Union[str, Any]=0.1 , SCREAMING_SNAKE_CASE_ : Optional[int]=0.1 , SCREAMING_SNAKE_CASE_ : List[str]=5_1_2 , SCREAMING_SNAKE_CASE_ : Optional[int]=6_4 , SCREAMING_SNAKE_CASE_ : Optional[int]=2 , SCREAMING_SNAKE_CASE_ : Optional[Any]=0.02 , SCREAMING_SNAKE_CASE_ : List[Any]=1e-1_2 , SCREAMING_SNAKE_CASE_ : int=0.1 , SCREAMING_SNAKE_CASE_ : int=0 , SCREAMING_SNAKE_CASE_ : Dict=2 , SCREAMING_SNAKE_CASE_ : List[str]=3 , SCREAMING_SNAKE_CASE_ : str=True , **SCREAMING_SNAKE_CASE_ : List[str] , ): super().__init__(pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) _a = vocab_size _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = hidden_act _a = intermediate_size _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = max_position_embeddings _a = max_relative_position _a = type_vocab_size _a = initializer_range _a = layer_norm_eps _a = classifier_dropout _a = use_cache
562
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int=1_3 , UpperCAmelCase__ : Optional[Any]=7 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Optional[int]=False , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Optional[int]=9_9 , UpperCAmelCase__ : Dict=3_2 , UpperCAmelCase__ : List[str]=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : Union[str, Any]=3_7 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=0.1 , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : int=5_1_2 , UpperCAmelCase__ : List[str]=1_6 , UpperCAmelCase__ : Optional[Any]=2 , UpperCAmelCase__ : List[Any]=0.02 , UpperCAmelCase__ : List[str]=3 , UpperCAmelCase__ : str=4 , UpperCAmelCase__ : List[Any]=None , ) -> Any: __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = scope def UpperCAmelCase_ ( self : int ) -> Dict: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_input_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = BioGptModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] , ) -> List[Any]: __SCREAMING_SNAKE_CASE = BioGptForCausalLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , *UpperCAmelCase__ : Optional[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE = BioGptModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() # create attention mask __SCREAMING_SNAKE_CASE = torch.ones(input_ids.shape , dtype=torch.long , device=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.seq_length // 2 __SCREAMING_SNAKE_CASE = 0 # first forward pass __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).to_tuple() # create hypothetical next token and extent to next_input_ids __SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids __SCREAMING_SNAKE_CASE = ids_tensor((1,) , UpperCAmelCase__ ).item() + 1 __SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = random_other_next_tokens # append to next input_ids and attn_mask __SCREAMING_SNAKE_CASE = torch.cat([input_ids, next_tokens] , dim=-1 ) __SCREAMING_SNAKE_CASE = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=UpperCAmelCase__ )] , dim=1 , ) # get two different outputs __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )["last_hidden_state"] __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )["last_hidden_state"] # select random slice __SCREAMING_SNAKE_CASE = ids_tensor((1,) , output_from_past.shape[-1] ).item() __SCREAMING_SNAKE_CASE = output_from_no_past[:, -1, random_slice_idx].detach() __SCREAMING_SNAKE_CASE = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-3 ) ) def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] , *UpperCAmelCase__ : Optional[Any] ) -> List[Any]: __SCREAMING_SNAKE_CASE = BioGptModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).eval() __SCREAMING_SNAKE_CASE = torch.ones(input_ids.shape , dtype=torch.long , device=UpperCAmelCase__ ) # first forward pass __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 3) , config.vocab_size ) __SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and __SCREAMING_SNAKE_CASE = torch.cat([input_ids, next_tokens] , dim=-1 ) __SCREAMING_SNAKE_CASE = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )["last_hidden_state"] __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )[ "last_hidden_state" ] # select random slice __SCREAMING_SNAKE_CASE = ids_tensor((1,) , output_from_past.shape[-1] ).item() __SCREAMING_SNAKE_CASE = output_from_no_past[:, -3:, random_slice_idx].detach() __SCREAMING_SNAKE_CASE = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-3 ) ) def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any , *UpperCAmelCase__ : Any , UpperCAmelCase__ : int=False ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = BioGptForCausalLM(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) if gradient_checkpointing: model.gradient_checkpointing_enable() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : str , *UpperCAmelCase__ : Optional[int] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = BioGptModel(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[Any] , *UpperCAmelCase__ : Dict ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = BioGptForTokenClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self : Optional[Any] ) -> str: __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ( __SCREAMING_SNAKE_CASE ) , ) = config_and_inputs __SCREAMING_SNAKE_CASE = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : Union[str, Any] = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case__ : Optional[int] = (BioGptForCausalLM,) if is_torch_available() else () snake_case__ : Tuple = ( { "feature-extraction": BioGptModel, "text-classification": BioGptForSequenceClassification, "text-generation": BioGptForCausalLM, "token-classification": BioGptForTokenClassification, "zero-shot": BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : Optional[Any] = False def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: __SCREAMING_SNAKE_CASE = BioGptModelTester(self ) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=3_7 ) def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[str] ) -> List[Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : int ) -> int: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ) -> Any: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*UpperCAmelCase__ , gradient_checkpointing=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Any: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*UpperCAmelCase__ ) @slow def UpperCAmelCase_ ( self : int ) -> List[str]: __SCREAMING_SNAKE_CASE = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) __SCREAMING_SNAKE_CASE = "left" # Define PAD Token = EOS Token = 50256 __SCREAMING_SNAKE_CASE = tokenizer.eos_token __SCREAMING_SNAKE_CASE = model.config.eos_token_id # use different length sentences to test batching __SCREAMING_SNAKE_CASE = [ "Hello, my dog is a little", "Today, I", ] __SCREAMING_SNAKE_CASE = tokenizer(UpperCAmelCase__ , return_tensors="pt" , padding=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = inputs["input_ids"].to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model.generate( input_ids=UpperCAmelCase__ , attention_mask=inputs["attention_mask"].to(UpperCAmelCase__ ) , ) __SCREAMING_SNAKE_CASE = tokenizer(sentences[0] , return_tensors="pt" ).input_ids.to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model.generate(input_ids=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = inputs_non_padded.shape[-1] - inputs["attention_mask"][-1].long().sum().cpu().item() __SCREAMING_SNAKE_CASE = tokenizer(sentences[1] , return_tensors="pt" ).input_ids.to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model.generate(input_ids=UpperCAmelCase__ , max_length=model.config.max_length - num_paddings ) __SCREAMING_SNAKE_CASE = tokenizer.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.decode(output_non_padded[0] , skip_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.decode(output_padded[0] , skip_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = [ "Hello, my dog is a little bit bigger than a little bit.", "Today, I have a good idea of how to use the information", ] self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , [non_padded_sentence, padded_sentence] ) @slow def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[int]: for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = BioGptModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Dict: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = 3 __SCREAMING_SNAKE_CASE = input_dict["input_ids"] __SCREAMING_SNAKE_CASE = input_ids.ne(1 ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __SCREAMING_SNAKE_CASE = BioGptForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def UpperCAmelCase_ ( self : List[Any] ) -> str: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE = 3 __SCREAMING_SNAKE_CASE = "multi_label_classification" __SCREAMING_SNAKE_CASE = input_dict["input_ids"] __SCREAMING_SNAKE_CASE = input_ids.ne(1 ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __SCREAMING_SNAKE_CASE = BioGptForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class UpperCamelCase_ ( unittest.TestCase): """simple docstring""" @slow def UpperCAmelCase_ ( self : int ) -> List[Any]: __SCREAMING_SNAKE_CASE = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) __SCREAMING_SNAKE_CASE = torch.tensor([[2, 4_8_0_5, 9, 6_5_6, 2_1]] ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ )[0] __SCREAMING_SNAKE_CASE = 4_2_3_8_4 __SCREAMING_SNAKE_CASE = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.tensor( [[[-9.5_236, -9.8_918, 10.4_557], [-11.0_469, -9.6_423, 8.1_022], [-8.8_664, -7.8_826, 5.5_325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase__ , atol=1E-4 ) ) @slow def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: __SCREAMING_SNAKE_CASE = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) __SCREAMING_SNAKE_CASE = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(UpperCAmelCase__ ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = tokenizer("COVID-19 is" , return_tensors="pt" ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model.generate( **UpperCAmelCase__ , min_length=1_0_0 , max_length=1_0_2_4 , num_beams=5 , early_stopping=UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = tokenizer.decode(output_ids[0] , skip_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = ( "COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the" " causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and" " territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK)," " and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and" " more than 800,000 deaths." ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
682
0
from ...configuration_utils import PretrainedConfig UpperCamelCase = { '''google/tapas-base-finetuned-sqa''': ( '''https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wtq''': ( '''https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wikisql-supervised''': ( '''https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json''' ), '''google/tapas-base-finetuned-tabfact''': ( '''https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json''' ), } class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : Tuple = "tapas" def __init__( self :Optional[int] , lowerCamelCase__ :List[str]=3_05_22 , lowerCamelCase__ :str=7_68 , lowerCamelCase__ :Tuple=12 , lowerCamelCase__ :List[Any]=12 , lowerCamelCase__ :Any=30_72 , lowerCamelCase__ :str="gelu" , lowerCamelCase__ :List[Any]=0.1 , lowerCamelCase__ :List[str]=0.1 , lowerCamelCase__ :Any=10_24 , lowerCamelCase__ :str=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , lowerCamelCase__ :Union[str, Any]=0.02 , lowerCamelCase__ :Tuple=1e-12 , lowerCamelCase__ :Any=0 , lowerCamelCase__ :Optional[Any]=10.0 , lowerCamelCase__ :Optional[Any]=0 , lowerCamelCase__ :Optional[int]=1.0 , lowerCamelCase__ :Dict=None , lowerCamelCase__ :Any=1.0 , lowerCamelCase__ :List[Any]=False , lowerCamelCase__ :Optional[int]=None , lowerCamelCase__ :str=1.0 , lowerCamelCase__ :Optional[Any]=1.0 , lowerCamelCase__ :List[str]=False , lowerCamelCase__ :List[str]=False , lowerCamelCase__ :Optional[Any]="ratio" , lowerCamelCase__ :str=None , lowerCamelCase__ :Optional[Any]=None , lowerCamelCase__ :Dict=64 , lowerCamelCase__ :int=32 , lowerCamelCase__ :Union[str, Any]=False , lowerCamelCase__ :Dict=True , lowerCamelCase__ :Optional[Any]=False , lowerCamelCase__ :Optional[int]=False , lowerCamelCase__ :Tuple=True , lowerCamelCase__ :str=False , lowerCamelCase__ :List[Any]=None , lowerCamelCase__ :Any=None , **lowerCamelCase__ :List[Any] , ): super().__init__(pad_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) UpperCamelCase__ :Optional[int] = vocab_size UpperCamelCase__ :int = hidden_size UpperCamelCase__ :int = num_hidden_layers UpperCamelCase__ :Tuple = num_attention_heads UpperCamelCase__ :Optional[int] = hidden_act UpperCamelCase__ :Optional[int] = intermediate_size UpperCamelCase__ :List[str] = hidden_dropout_prob UpperCamelCase__ :Union[str, Any] = attention_probs_dropout_prob UpperCamelCase__ :Optional[Any] = max_position_embeddings UpperCamelCase__ :Dict = type_vocab_sizes UpperCamelCase__ :Any = initializer_range UpperCamelCase__ :Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters UpperCamelCase__ :Any = positive_label_weight UpperCamelCase__ :Any = num_aggregation_labels UpperCamelCase__ :Union[str, Any] = aggregation_loss_weight UpperCamelCase__ :Tuple = use_answer_as_supervision UpperCamelCase__ :int = answer_loss_importance UpperCamelCase__ :str = use_normalized_answer_loss UpperCamelCase__ :Optional[Any] = huber_loss_delta UpperCamelCase__ :str = temperature UpperCamelCase__ :str = aggregation_temperature UpperCamelCase__ :Union[str, Any] = use_gumbel_for_cells UpperCamelCase__ :str = use_gumbel_for_aggregation UpperCamelCase__ :List[Any] = average_approximation_function UpperCamelCase__ :Tuple = cell_selection_preference UpperCamelCase__ :Optional[int] = answer_loss_cutoff UpperCamelCase__ :Optional[Any] = max_num_rows UpperCamelCase__ :Dict = max_num_columns UpperCamelCase__ :Optional[int] = average_logits_per_cell UpperCamelCase__ :List[str] = select_one_column UpperCamelCase__ :Dict = allow_empty_column_selection UpperCamelCase__ :List[str] = init_cell_selection_weights_to_zero UpperCamelCase__ :Tuple = reset_position_index_per_cell UpperCamelCase__ :int = disable_per_token_loss # Aggregation hyperparameters UpperCamelCase__ :Any = aggregation_labels UpperCamelCase__ :Tuple = no_aggregation_label_index if isinstance(self.aggregation_labels , UpperCAmelCase__ ): UpperCamelCase__ :Union[str, Any] = {int(UpperCAmelCase__ ): v for k, v in aggregation_labels.items()}
45
"""simple docstring""" import os import pytest from attr import dataclass a__ : int = '''us-east-1''' # defaults region @dataclass class UpperCamelCase_ : """simple docstring""" snake_case__ : str snake_case__ : Optional[Any] = "arn:aws:iam::558105141721:role/sagemaker_execution_role" snake_case__ : Optional[Any] = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 500, "save_steps": 5500, } snake_case__ : Tuple = {**hyperparameters, "max_steps": 1000} @property def UpperCAmelCase_ ( self : Any ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCAmelCase_ ( self : int ) -> str: return F"""{self.framework}-transfromers-test""" @property def UpperCAmelCase_ ( self : List[Any] ) -> str: return F"""./tests/sagemaker/scripts/{self.framework}""" @property def UpperCAmelCase_ ( self : Any ) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="class" ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = SageMakerTestEnvironment(framework=request.cls.framework )
682
0
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run lowerCamelCase__ = True except (ImportError, AttributeError): lowerCamelCase__ = object def lowercase__ ( *lowercase_ ,**lowercase_ ) -> str: """simple docstring""" pass lowerCamelCase__ = False lowerCamelCase__ = logging.get_logger("transformers-cli/serving") def lowercase__ ( lowercase_ ) -> str: """simple docstring""" _UpperCamelCase : int = pipeline( task=args.task ,model=args.model if args.model else None ,config=args.config ,tokenizer=args.tokenizer ,device=args.device ,) return ServeCommand(lowerCAmelCase_ ,args.host ,args.port ,args.workers ) class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :dict class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :List[str] SCREAMING_SNAKE_CASE__ :Optional[List[int]] class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :str class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ :Any class __SCREAMING_SNAKE_CASE ( _UpperCamelCase ): '''simple docstring''' @staticmethod def __SCREAMING_SNAKE_CASE ( __a : ArgumentParser ) -> Optional[int]: _UpperCamelCase : List[Any] = parser.add_parser( "serve" , help="CLI tool to run inference requests through REST and GraphQL endpoints." ) serve_parser.add_argument( "--task" , type=UpperCAmelCase__ , choices=get_supported_tasks() , help="The task to run the pipeline on" , ) serve_parser.add_argument("--host" , type=UpperCAmelCase__ , default="localhost" , help="Interface the server will listen on." ) serve_parser.add_argument("--port" , type=UpperCAmelCase__ , default=8888 , help="Port the serving will listen to." ) serve_parser.add_argument("--workers" , type=UpperCAmelCase__ , default=1 , help="Number of http workers" ) serve_parser.add_argument("--model" , type=UpperCAmelCase__ , help="Model's name or path to stored model." ) serve_parser.add_argument("--config" , type=UpperCAmelCase__ , help="Model's config name or path to stored model." ) serve_parser.add_argument("--tokenizer" , type=UpperCAmelCase__ , help="Tokenizer name to use." ) serve_parser.add_argument( "--device" , type=UpperCAmelCase__ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) serve_parser.set_defaults(func=UpperCAmelCase__ ) def __init__( self : Optional[Any] , __a : Pipeline , __a : str , __a : int , __a : int ) -> Optional[Any]: _UpperCamelCase : List[Any] = pipeline _UpperCamelCase : int = host _UpperCamelCase : Tuple = port _UpperCamelCase : Dict = workers if not _serve_dependencies_installed: raise RuntimeError( "Using serve command requires FastAPI and uvicorn. " "Please install transformers with [serving]: pip install \"transformers[serving]\"." "Or install FastAPI and uvicorn separately." ) else: logger.info(F'''Serving model over {host}:{port}''' ) _UpperCamelCase : Optional[Any] = FastAPI( routes=[ APIRoute( "/" , self.model_info , response_model=UpperCAmelCase__ , response_class=UpperCAmelCase__ , methods=["GET"] , ), APIRoute( "/tokenize" , self.tokenize , response_model=UpperCAmelCase__ , response_class=UpperCAmelCase__ , methods=["POST"] , ), APIRoute( "/detokenize" , self.detokenize , response_model=UpperCAmelCase__ , response_class=UpperCAmelCase__ , methods=["POST"] , ), APIRoute( "/forward" , self.forward , response_model=UpperCAmelCase__ , response_class=UpperCAmelCase__ , methods=["POST"] , ), ] , timeout=600 , ) def __SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[int]: run(self._app , host=self.host , port=self.port , workers=self.workers ) def __SCREAMING_SNAKE_CASE ( self : List[str] ) -> Union[str, Any]: return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def __SCREAMING_SNAKE_CASE ( self : Optional[Any] , __a : str = Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) , __a : bool = Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) ) -> Union[str, Any]: try: _UpperCamelCase : Dict = self._pipeline.tokenizer.tokenize(UpperCAmelCase__ ) if return_ids: _UpperCamelCase : int = self._pipeline.tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) return ServeTokenizeResult(tokens=UpperCAmelCase__ , tokens_ids=UpperCAmelCase__ ) else: return ServeTokenizeResult(tokens=UpperCAmelCase__ ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(UpperCAmelCase__ )} ) def __SCREAMING_SNAKE_CASE ( self : Any , __a : List[int] = Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) , __a : bool = Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) , __a : bool = Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) , ) -> List[str]: try: _UpperCamelCase : Tuple = self._pipeline.tokenizer.decode(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return ServeDeTokenizeResult(model="" , text=UpperCAmelCase__ ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(UpperCAmelCase__ )} ) async def __SCREAMING_SNAKE_CASE ( self : Any , __a : int=Body(UpperCAmelCase__ , embed=UpperCAmelCase__ ) ) -> Tuple: # Check we don't have empty string if len(UpperCAmelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _UpperCamelCase : Tuple = self._pipeline(UpperCAmelCase__ ) return ServeForwardResult(output=UpperCAmelCase__ ) except Exception as e: raise HTTPException(500 , {"error": str(UpperCAmelCase__ )} )
624
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging a__ : Any = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Any , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Union[str, Any] ) -> Any: warnings.warn( "`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` " "instead." , UpperCAmelCase__ , ) super().__init__(args=UpperCAmelCase__ , **UpperCAmelCase__ )
682
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , __a , __a=7 , __a=3 , __a=18 , __a=30 , __a=4_00 , __a=True , __a=None , __a=True , ): __lowerCAmelCase = size if size is not None else {"height": 18, "width": 18} __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = image_size __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = apply_ocr def snake_case ( self ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class _UpperCamelCase ( lowerCAmelCase__ ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Tuple =LayoutLMvaImageProcessor if is_pytesseract_available() else None def snake_case ( self ): __lowerCAmelCase = LayoutLMvaImageProcessingTester(self ) @property def snake_case ( self ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case ( self ): __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , "do_resize" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , "size" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , "apply_ocr" ) ) def snake_case ( self ): __lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) __lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def snake_case ( self ): pass def snake_case ( self ): # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched __lowerCAmelCase = image_processing(UpperCAmelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def snake_case ( self ): # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __lowerCAmelCase = image_processing(UpperCAmelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def snake_case ( self ): # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __lowerCAmelCase = image_processing(UpperCAmelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def snake_case ( self ): # with apply_OCR = True __lowerCAmelCase = LayoutLMvaImageProcessor() from datasets import load_dataset __lowerCAmelCase = load_dataset("hf-internal-testing/fixtures_docvqa" , split="test" ) __lowerCAmelCase = Image.open(ds[0]["file"] ).convert("RGB" ) __lowerCAmelCase = image_processing(UpperCAmelCase__ , return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __lowerCAmelCase = [["11:14", "to", "11:39", "a.m", "11:39", "to", "11:44", "a.m.", "11:44", "a.m.", "to", "12:25", "p.m.", "12:25", "to", "12:58", "p.m.", "12:58", "to", "4:00", "p.m.", "2:00", "to", "5:00", "p.m.", "Coffee", "Break", "Coffee", "will", "be", "served", "for", "men", "and", "women", "in", "the", "lobby", "adjacent", "to", "exhibit", "area.", "Please", "move", "into", "exhibit", "area.", "(Exhibits", "Open)", "TRRF", "GENERAL", "SESSION", "(PART", "|)", "Presiding:", "Lee", "A.", "Waller", "TRRF", "Vice", "President", "“Introductory", "Remarks”", "Lee", "A.", "Waller,", "TRRF", "Vice", "Presi-", "dent", "Individual", "Interviews", "with", "TRRF", "Public", "Board", "Members", "and", "Sci-", "entific", "Advisory", "Council", "Mem-", "bers", "Conducted", "by", "TRRF", "Treasurer", "Philip", "G.", "Kuehn", "to", "get", "answers", "which", "the", "public", "refrigerated", "warehousing", "industry", "is", "looking", "for.", "Plus", "questions", "from", "the", "floor.", "Dr.", "Emil", "M.", "Mrak,", "University", "of", "Cal-", "ifornia,", "Chairman,", "TRRF", "Board;", "Sam", "R.", "Cecil,", "University", "of", "Georgia", "College", "of", "Agriculture;", "Dr.", "Stanley", "Charm,", "Tufts", "University", "School", "of", "Medicine;", "Dr.", "Robert", "H.", "Cotton,", "ITT", "Continental", "Baking", "Company;", "Dr.", "Owen", "Fennema,", "University", "of", "Wis-", "consin;", "Dr.", "Robert", "E.", "Hardenburg,", "USDA.", "Questions", "and", "Answers", "Exhibits", "Open", "Capt.", "Jack", "Stoney", "Room", "TRRF", "Scientific", "Advisory", "Council", "Meeting", "Ballroom", "Foyer"]] # noqa: E231 __lowerCAmelCase = [[[1_41, 57, 2_14, 69], [2_28, 58, 2_52, 69], [1_41, 75, 2_16, 88], [2_30, 79, 2_80, 88], [1_42, 2_60, 2_18, 2_73], [2_30, 2_61, 2_55, 2_73], [1_43, 2_79, 2_18, 2_90], [2_31, 2_82, 2_90, 2_91], [1_43, 3_42, 2_18, 3_54], [2_31, 3_45, 2_89, 3_55], [2_02, 3_62, 2_27, 3_73], [1_43, 3_79, 2_20, 3_92], [2_31, 3_82, 2_91, 3_94], [1_44, 7_14, 2_20, 7_26], [2_31, 7_15, 2_56, 7_26], [1_44, 7_32, 2_20, 7_45], [2_32, 7_36, 2_91, 7_47], [1_44, 7_69, 2_18, 7_82], [2_31, 7_70, 2_56, 7_82], [1_41, 7_88, 2_02, 8_01], [2_15, 7_91, 2_74, 8_04], [1_43, 8_26, 2_04, 8_38], [2_15, 8_26, 2_40, 8_38], [1_42, 8_44, 2_02, 8_57], [2_15, 8_47, 2_74, 8_59], [3_34, 57, 4_27, 69], [4_40, 57, 5_22, 69], [3_69, 75, 4_61, 88], [4_69, 75, 5_16, 88], [5_28, 76, 5_62, 88], [5_70, 76, 6_67, 88], [6_75, 75, 7_11, 87], [7_21, 79, 7_78, 88], [7_89, 75, 8_40, 88], [3_69, 97, 4_70, 1_07], [4_84, 94, 5_07, 1_06], [5_18, 94, 5_62, 1_07], [5_76, 94, 6_55, 1_10], [6_68, 94, 7_92, 1_09], [8_04, 95, 8_29, 1_07], [3_69, 1_13, 4_65, 1_25], [4_77, 1_16, 5_47, 1_25], [5_62, 1_13, 6_58, 1_25], [6_71, 1_16, 7_48, 1_25], [7_61, 1_13, 8_11, 1_25], [3_69, 1_31, 4_65, 1_43], [4_77, 1_33, 5_48, 1_43], [5_63, 1_30, 6_98, 1_45], [7_10, 1_30, 8_02, 1_46], [3_36, 1_71, 4_12, 1_83], [4_23, 1_71, 5_72, 1_83], [5_82, 1_70, 7_16, 1_84], [7_28, 1_71, 8_17, 1_87], [8_29, 1_71, 8_44, 1_86], [3_38, 1_97, 4_82, 2_12], [5_07, 1_96, 5_57, 2_09], [5_69, 1_96, 5_95, 2_08], [6_10, 1_96, 7_02, 2_09], [5_05, 2_14, 5_83, 2_26], [5_95, 2_14, 6_56, 2_27], [6_70, 2_15, 8_07, 2_27], [3_35, 2_59, 5_43, 2_74], [5_56, 2_59, 7_08, 2_72], [3_72, 2_79, 4_22, 2_91], [4_35, 2_79, 4_60, 2_91], [4_74, 2_79, 5_74, 2_92], [5_87, 2_78, 6_64, 2_91], [6_76, 2_78, 7_38, 2_91], [7_51, 2_79, 8_34, 2_91], [3_72, 2_98, 4_34, 3_10], [3_35, 3_41, 4_83, 3_54], [4_97, 3_41, 6_55, 3_54], [6_67, 3_41, 7_28, 3_54], [7_40, 3_41, 8_25, 3_54], [3_35, 3_60, 4_30, 3_72], [4_42, 3_60, 5_34, 3_72], [5_45, 3_59, 6_87, 3_72], [6_97, 3_60, 7_54, 3_72], [7_65, 3_60, 8_23, 3_73], [3_34, 3_78, 4_28, 3_91], [4_40, 3_78, 5_77, 3_94], [5_90, 3_78, 7_05, 3_91], [7_20, 3_78, 8_01, 3_91], [3_34, 3_97, 4_00, 4_09], [3_70, 4_16, 5_29, 4_29], [5_44, 4_16, 5_76, 4_32], [5_87, 4_16, 6_65, 4_28], [6_77, 4_16, 8_14, 4_29], [3_72, 4_35, 4_52, 4_50], [4_65, 4_34, 4_95, 4_47], [5_11, 4_34, 6_00, 4_47], [6_11, 4_36, 6_37, 4_47], [6_49, 4_36, 6_94, 4_51], [7_05, 4_38, 8_24, 4_47], [3_69, 4_53, 4_52, 4_66], [4_64, 4_54, 5_09, 4_66], [5_22, 4_53, 6_11, 4_69], [6_25, 4_53, 7_92, 4_69], [3_70, 4_72, 5_56, 4_88], [5_70, 4_72, 6_84, 4_87], [6_97, 4_72, 7_18, 4_85], [7_32, 4_72, 8_35, 4_88], [3_69, 4_90, 4_11, 5_03], [4_25, 4_90, 4_84, 5_03], [4_96, 4_90, 6_35, 5_06], [6_45, 4_90, 7_07, 5_03], [7_18, 4_91, 7_61, 5_03], [7_71, 4_90, 8_40, 5_03], [3_36, 5_10, 3_74, 5_21], [3_88, 5_10, 4_47, 5_22], [4_60, 5_10, 4_89, 5_21], [5_03, 5_10, 5_80, 5_22], [5_92, 5_09, 7_36, 5_25], [7_45, 5_09, 7_70, 5_22], [7_81, 5_09, 8_40, 5_22], [3_38, 5_28, 4_34, 5_41], [4_48, 5_28, 5_96, 5_41], [6_09, 5_27, 6_87, 5_40], [7_00, 5_28, 7_92, 5_41], [3_36, 5_46, 3_97, 5_59], [4_07, 5_46, 4_31, 5_59], [4_43, 5_46, 5_25, 5_60], [5_37, 5_46, 6_80, 5_62], [6_88, 5_46, 7_14, 5_59], [7_22, 5_46, 8_37, 5_62], [3_36, 5_65, 4_49, 5_81], [4_61, 5_65, 4_85, 5_77], [4_97, 5_65, 6_65, 5_81], [6_81, 5_65, 7_18, 5_77], [7_32, 5_65, 8_37, 5_80], [3_37, 5_84, 4_38, 5_97], [4_52, 5_83, 5_21, 5_96], [5_35, 5_84, 6_77, 5_99], [6_90, 5_83, 7_87, 5_96], [8_01, 5_83, 8_25, 5_96], [3_38, 6_02, 4_78, 6_15], [4_92, 6_02, 5_30, 6_14], [5_43, 6_02, 6_38, 6_15], [6_50, 6_02, 6_76, 6_14], [6_88, 6_02, 7_88, 6_15], [8_02, 6_02, 8_43, 6_14], [3_37, 6_21, 5_02, 6_33], [5_16, 6_21, 6_15, 6_37], [6_29, 6_21, 7_74, 6_36], [7_89, 6_21, 8_27, 6_33], [3_37, 6_39, 4_18, 6_52], [4_32, 6_40, 5_71, 6_53], [5_87, 6_39, 7_31, 6_55], [7_43, 6_39, 7_69, 6_52], [7_80, 6_39, 8_41, 6_52], [3_38, 6_58, 4_40, 6_73], [4_55, 6_58, 4_91, 6_70], [5_08, 6_58, 6_02, 6_71], [6_16, 6_58, 6_38, 6_70], [6_54, 6_58, 8_35, 6_74], [3_37, 6_77, 4_29, 6_89], [3_37, 7_14, 4_82, 7_26], [4_95, 7_14, 5_48, 7_26], [5_61, 7_14, 6_83, 7_26], [3_38, 7_70, 4_61, 7_82], [4_74, 7_69, 5_54, 7_85], [4_89, 7_88, 5_62, 8_03], [5_76, 7_88, 6_43, 8_01], [6_56, 7_87, 7_51, 8_04], [7_64, 7_88, 8_44, 8_01], [3_34, 8_25, 4_21, 8_38], [4_30, 8_24, 5_74, 8_38], [5_84, 8_24, 7_23, 8_41], [3_35, 8_44, 4_50, 8_57], [4_64, 8_43, 5_83, 8_60], [6_28, 8_62, 7_55, 8_75], [7_69, 8_61, 8_48, 8_78]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False __lowerCAmelCase = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) __lowerCAmelCase = image_processing(UpperCAmelCase__ , return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) )
636
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if collection == []: return [] # get some information about the collection __SCREAMING_SNAKE_CASE = len(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = max(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = min(lowerCAmelCase_ ) # create the counting array __SCREAMING_SNAKE_CASE = coll_max + 1 - coll_min __SCREAMING_SNAKE_CASE = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = counting_arr[i] + counting_arr[i - 1] # create the output collection __SCREAMING_SNAKE_CASE = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCAmelCase_ ) ): __SCREAMING_SNAKE_CASE = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' return "".join([chr(lowerCAmelCase_ ) for i in counting_sort([ord(lowerCAmelCase_ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string('''thisisthestring''') == "eghhiiinrsssttt" a__ : Dict = input('''Enter numbers separated by a comma:\n''').strip() a__ : Optional[Any] = [int(item) for item in user_input.split(''',''')] print(counting_sort(unsorted))
682
0
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class _a ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) SCREAMING_SNAKE_CASE : Dict = AutoTokenizer.from_pretrained('xlm-roberta-base' ) SCREAMING_SNAKE_CASE : Optional[Any] = 'The dog is cute and lives in the garden house' SCREAMING_SNAKE_CASE : Optional[int] = jnp.array([tokenizer.encode(UpperCAmelCase__ )] ) SCREAMING_SNAKE_CASE : Optional[Any] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.array( [[-0.01_01, 0.12_18, -0.08_03, 0.08_01, 0.13_27, 0.07_76, -0.12_15, 0.23_83, 0.33_38, 0.31_06, 0.03_00, 0.02_52]] ) SCREAMING_SNAKE_CASE : Optional[int] = model(UpperCAmelCase__ )['last_hidden_state'] self.assertEqual(output.shape, UpperCAmelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1], UpperCAmelCase__, atol=1E-3 ) )
28
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a__ : Tuple = { '''configuration_rag''': ['''RagConfig'''], '''retrieval_rag''': ['''RagRetriever'''], '''tokenization_rag''': ['''RagTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = [ '''RagModel''', '''RagPreTrainedModel''', '''RagSequenceForGeneration''', '''RagTokenForGeneration''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = [ '''TFRagModel''', '''TFRagPreTrainedModel''', '''TFRagSequenceForGeneration''', '''TFRagTokenForGeneration''', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys a__ : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
682
0
"""simple docstring""" import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase ) -> int: '''simple docstring''' _lowerCamelCase : Union[str, Any] = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg" _lowerCamelCase : Optional[Any] = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ).convert("RGB" ) _lowerCamelCase : Any = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3) , (0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1) ), ] ) _lowerCamelCase : Dict = transform(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) return image def lowerCamelCase_( _lowerCamelCase ) -> Any: '''simple docstring''' if "visual_encoder" in key: _lowerCamelCase : Tuple = re.sub("visual_encoder*" , "vision_model.encoder" , lowerCAmelCase_ ) if "blocks" in key: _lowerCamelCase : Dict = re.sub(R"blocks" , "layers" , lowerCAmelCase_ ) if "attn" in key: _lowerCamelCase : List[str] = re.sub(R"attn" , "self_attn" , lowerCAmelCase_ ) if "norm1" in key: _lowerCamelCase : int = re.sub(R"norm1" , "layer_norm1" , lowerCAmelCase_ ) if "norm2" in key: _lowerCamelCase : Dict = re.sub(R"norm2" , "layer_norm2" , lowerCAmelCase_ ) if "encoder.norm" in key: _lowerCamelCase : int = re.sub(R"encoder.norm" , "post_layernorm" , lowerCAmelCase_ ) if "encoder.patch_embed.proj" in key: _lowerCamelCase : List[Any] = re.sub(R"encoder.patch_embed.proj" , "embeddings.patch_embedding" , lowerCAmelCase_ ) if "encoder.pos_embed" in key: _lowerCamelCase : str = re.sub(R"encoder.pos_embed" , "embeddings.position_embedding" , lowerCAmelCase_ ) if "encoder.cls_token" in key: _lowerCamelCase : List[str] = re.sub(R"encoder.cls_token" , "embeddings.class_embedding" , lowerCAmelCase_ ) if "self_attn" in key: _lowerCamelCase : Union[str, Any] = re.sub(R"self_attn.proj" , "self_attn.projection" , lowerCAmelCase_ ) return key @torch.no_grad() def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase=None ) -> Dict: '''simple docstring''' if config_path is not None: _lowerCamelCase : int = BlipConfig.from_pretrained(lowerCAmelCase_ ) else: _lowerCamelCase : Optional[int] = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) _lowerCamelCase : Dict = BlipForConditionalGeneration(lowerCAmelCase_ ).eval() _lowerCamelCase : Optional[Any] = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth" _lowerCamelCase : Optional[int] = blip_decoder(pretrained=lowerCAmelCase_ , image_size=384 , vit="base" ) _lowerCamelCase : Optional[Any] = pt_model.eval() _lowerCamelCase : List[str] = pt_model.state_dict() for key in modified_state_dict.copy(): _lowerCamelCase : str = modified_state_dict.pop(lowerCAmelCase_ ) _lowerCamelCase : Dict = rename_key(lowerCAmelCase_ ) _lowerCamelCase : str = value hf_model.load_state_dict(lowerCAmelCase_ ) _lowerCamelCase : Optional[int] = 384 _lowerCamelCase : Optional[int] = load_demo_image(image_size=lowerCAmelCase_ , device="cpu" ) _lowerCamelCase : int = BertTokenizer.from_pretrained("bert-base-uncased" ) _lowerCamelCase : List[str] = tokenizer(["a picture of"] ).input_ids _lowerCamelCase : Optional[Any] = hf_model.generate(lowerCAmelCase_ , lowerCAmelCase_ ) assert out[0].tolist() == [30522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] _lowerCamelCase : Optional[Any] = hf_model.generate(lowerCAmelCase_ ) assert out[0].tolist() == [30522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(lowerCAmelCase_ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' _lowerCamelCase : Tuple = ( "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth" ) _lowerCamelCase : Optional[int] = blip_vqa(pretrained=lowerCAmelCase_ , image_size=lowerCAmelCase_ , vit="base" ) vqa_model.eval() _lowerCamelCase : Tuple = vqa_model.state_dict() for key in modified_state_dict.copy(): _lowerCamelCase : Tuple = modified_state_dict.pop(lowerCAmelCase_ ) _lowerCamelCase : Tuple = rename_key(lowerCAmelCase_ ) _lowerCamelCase : int = value _lowerCamelCase : str = BlipForQuestionAnswering(lowerCAmelCase_ ) hf_vqa_model.load_state_dict(lowerCAmelCase_ ) _lowerCamelCase : Dict = ["How many dogs are in this image?"] _lowerCamelCase : List[Any] = tokenizer(lowerCAmelCase_ , return_tensors="pt" ).input_ids _lowerCamelCase : int = hf_vqa_model.generate(lowerCAmelCase_ , lowerCAmelCase_ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + "_vqa" ) _lowerCamelCase : Any = "https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth" _lowerCamelCase : Optional[Any] = blip_itm(pretrained=lowerCAmelCase_ , image_size=lowerCAmelCase_ , vit="base" ) itm_model.eval() _lowerCamelCase : Dict = itm_model.state_dict() for key in modified_state_dict.copy(): _lowerCamelCase : Any = modified_state_dict.pop(lowerCAmelCase_ ) _lowerCamelCase : Tuple = rename_key(lowerCAmelCase_ ) _lowerCamelCase : Optional[Any] = value _lowerCamelCase : List[str] = BlipForImageTextRetrieval(lowerCAmelCase_ ) _lowerCamelCase : List[str] = ["A picture of a woman with a dog sitting in a beach"] _lowerCamelCase : Union[str, Any] = tokenizer( lowerCAmelCase_ , return_tensors="pt" , padding="max_length" , truncation=lowerCAmelCase_ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(lowerCAmelCase_ ) hf_itm_model.eval() _lowerCamelCase : Optional[Any] = hf_itm_model(lowerCAmelCase_ , lowerCAmelCase_ , use_itm_head=lowerCAmelCase_ ) _lowerCamelCase : List[Any] = hf_itm_model(lowerCAmelCase_ , lowerCAmelCase_ , use_itm_head=lowerCAmelCase_ ) assert out[0].item() == 0.2_1_1_0_6_8_7_4_9_4_2_7_7_9_5_4 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5_6_9_8_8_4_5_3_8_6_5_0_5_1_2_7 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + "_itm" ) if __name__ == "__main__": _lowerCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') _lowerCAmelCase : Dict = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
46
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a__ : List[str] = logging.get_logger(__name__) a__ : str = { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/config.json''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/config.json''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json''' ), } class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Any = "xlm-roberta" def __init__( self : int , UpperCAmelCase__ : Union[str, Any]=3_0_5_2_2 , UpperCAmelCase__ : Optional[Any]=7_6_8 , UpperCAmelCase__ : Optional[int]=1_2 , UpperCAmelCase__ : Tuple=1_2 , UpperCAmelCase__ : str=3_0_7_2 , UpperCAmelCase__ : int="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : Optional[int]=5_1_2 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : int=0.02 , UpperCAmelCase__ : Optional[int]=1E-12 , UpperCAmelCase__ : Any=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : Any="absolute" , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : List[Any]=None , **UpperCAmelCase__ : int , ) -> Tuple: super().__init__(pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = position_embedding_type __SCREAMING_SNAKE_CASE = use_cache __SCREAMING_SNAKE_CASE = classifier_dropout class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" @property def UpperCAmelCase_ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __SCREAMING_SNAKE_CASE = {0: "batch", 1: "choice", 2: "sequence"} else: __SCREAMING_SNAKE_CASE = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
682
0
from __future__ import annotations import math def A__ ( SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Any: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True UpperCAmelCase_ = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def A__ ( SCREAMING_SNAKE_CASE_ : List[str] ) -> str: """simple docstring""" if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError('''n must be an integer''' ) if n <= 0: raise ValueError('''n must be >= 0''' ) _UpperCAmelCase = [] for num in range(len(lowerCAmelCase_ ) ): _UpperCAmelCase = 0 while 2 * i * i <= odd_composites[num]: _UpperCAmelCase = odd_composites[num] - 2 * i * i if is_prime(lowerCAmelCase_ ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(lowerCAmelCase_ ) == n: return list_nums return [] def A__ ( ) -> int: """simple docstring""" return compute_nums(1 )[0] if __name__ == "__main__": print(f'''{solution() = }''')
32
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = checkpoints.load_tax_checkpoint(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = flatten_dict(lowerCAmelCase_ ) return flax_params def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = { "token_embedder": "embeddings", "encoder_norm": "layernorm", "kernel": "weight", ".out": ".output", "scale": "weight", "embedders_0.pos_embedding": "row_embedder.weight", "embedders_1.pos_embedding": "column_embedder.weight", } __SCREAMING_SNAKE_CASE = { "query": "attention.query", "key": "attention.key", "value": "attention.value", "output.dense": "output", "encoder_decoder_attention.o": "encoder_decoder_attention.attention.o", "pre_self_attention_layer_norm": "self_attention.layer_norm", "pre_cross_attention_layer_norm": "encoder_decoder_attention.layer_norm", "mlp.": "mlp.DenseReluDense.", "pre_mlp_layer_norm": "mlp.layer_norm", "self_attention.o": "self_attention.attention.o", "decoder.embeddings.embedding": "decoder.embed_tokens.weight", "decoder.relpos_bias.rel_embedding": "decoder.layer.0.self_attention.attention.relative_attention_bias.weight", "decoder.decoder_norm.weight": "decoder.final_layer_norm.weight", "decoder.logits_dense.weight": "decoder.lm_head.weight", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key __SCREAMING_SNAKE_CASE = ".".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): __SCREAMING_SNAKE_CASE = new_key.replace(lowerCAmelCase_ , lowerCAmelCase_ ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): __SCREAMING_SNAKE_CASE = new_key.replace(lowerCAmelCase_ , lowerCAmelCase_ ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number __SCREAMING_SNAKE_CASE = re.sub(R"layers_(\d+)" , R"layer.\1" , lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = new_key.replace("encoder" , "encoder.encoder" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number __SCREAMING_SNAKE_CASE = re.sub(R"layers_(\d+)" , R"layer.\1" , lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = flax_dict[key] __SCREAMING_SNAKE_CASE = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): __SCREAMING_SNAKE_CASE = torch.from_numpy(converted_dict[key].T ) else: __SCREAMING_SNAKE_CASE = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False , lowerCAmelCase_=False ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_flax_param(lowerCAmelCase_ ) if not use_large: __SCREAMING_SNAKE_CASE = PixaStructVisionConfig() __SCREAMING_SNAKE_CASE = PixaStructTextConfig() else: __SCREAMING_SNAKE_CASE = PixaStructVisionConfig( hidden_size=1536 , d_ff=3968 , num_attention_heads=24 , num_hidden_layers=18 ) __SCREAMING_SNAKE_CASE = PixaStructTextConfig(hidden_size=1536 , d_ff=3968 , num_heads=24 , num_layers=18 ) __SCREAMING_SNAKE_CASE = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = PixaStructForConditionalGeneration(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = rename_and_convert_flax_params(lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("ybelkada/test-pix2struct-tokenizer" ) __SCREAMING_SNAKE_CASE = PixaStructImageProcessor() __SCREAMING_SNAKE_CASE = PixaStructProcessor(image_processor=lowerCAmelCase_ , tokenizer=lowerCAmelCase_ ) if use_large: __SCREAMING_SNAKE_CASE = 4096 __SCREAMING_SNAKE_CASE = True # mkdir if needed os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) print("Model saved in {}".format(lowerCAmelCase_ ) ) if __name__ == "__main__": a__ : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') a__ : Optional[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
682
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase__ ( self : Optional[Any] ): '''simple docstring''' lowercase__ = tempfile.mkdtemp() lowercase__ = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file, '''w''', encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) lowercase__ = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48145466, 0.4578275, 0.40821073], '''image_std''': [0.26862954, 0.26130258, 0.27577711], } lowercase__ = os.path.join(self.tmpdirname, UpperCAmelCase__ ) with open(self.image_processor_file, '''w''', encoding='''utf-8''' ) as fp: json.dump(UpperCAmelCase__, UpperCAmelCase__ ) def lowercase__ ( self : int, **lowerCamelCase : Dict ): '''simple docstring''' return BertTokenizer.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def lowercase__ ( self : Tuple, **lowerCamelCase : Optional[int] ): '''simple docstring''' return BertTokenizerFast.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def lowercase__ ( self : Union[str, Any], **lowerCamelCase : Optional[int] ): '''simple docstring''' return EfficientNetImageProcessor.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def lowercase__ ( self : Dict ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowercase__ ( self : Dict ): '''simple docstring''' lowercase__ = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] lowercase__ = [Image.fromarray(np.moveaxis(UpperCAmelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def lowercase__ ( self : Optional[Any] ): '''simple docstring''' lowercase__ = self.get_tokenizer() lowercase__ = self.get_rust_tokenizer() lowercase__ = self.get_image_processor() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) processor_slow.save_pretrained(self.tmpdirname ) lowercase__ = AlignProcessor.from_pretrained(self.tmpdirname, use_fast=UpperCAmelCase__ ) lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) processor_fast.save_pretrained(self.tmpdirname ) lowercase__ = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab(), tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab(), tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab(), tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer, UpperCAmelCase__ ) self.assertIsInstance(processor_fast.tokenizer, UpperCAmelCase__ ) self.assertEqual(processor_slow.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor, UpperCAmelCase__ ) self.assertIsInstance(processor_fast.image_processor, UpperCAmelCase__ ) def lowercase__ ( self : List[Any] ): '''simple docstring''' lowercase__ = AlignProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowercase__ = self.get_tokenizer(bos_token='''(BOS)''', eos_token='''(EOS)''' ) lowercase__ = self.get_image_processor(do_normalize=UpperCAmelCase__, padding_value=1.0 ) lowercase__ = AlignProcessor.from_pretrained( self.tmpdirname, bos_token='''(BOS)''', eos_token='''(EOS)''', do_normalize=UpperCAmelCase__, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, UpperCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCAmelCase__ ) def lowercase__ ( self : List[Any] ): '''simple docstring''' lowercase__ = self.get_image_processor() lowercase__ = self.get_tokenizer() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) lowercase__ = self.prepare_image_inputs() lowercase__ = image_processor(UpperCAmelCase__, return_tensors='''np''' ) lowercase__ = processor(images=UpperCAmelCase__, return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum(), input_processor[key].sum(), delta=1E-2 ) def lowercase__ ( self : Tuple ): '''simple docstring''' lowercase__ = self.get_image_processor() lowercase__ = self.get_tokenizer() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) lowercase__ = '''lower newer''' lowercase__ = processor(text=UpperCAmelCase__ ) lowercase__ = tokenizer(UpperCAmelCase__, padding='''max_length''', max_length=64 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def lowercase__ ( self : Any ): '''simple docstring''' lowercase__ = self.get_image_processor() lowercase__ = self.get_tokenizer() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) lowercase__ = '''lower newer''' lowercase__ = self.prepare_image_inputs() lowercase__ = processor(text=UpperCAmelCase__, images=UpperCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase__ ): processor() def lowercase__ ( self : List[str] ): '''simple docstring''' lowercase__ = self.get_image_processor() lowercase__ = self.get_tokenizer() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) lowercase__ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowercase__ = processor.batch_decode(UpperCAmelCase__ ) lowercase__ = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__, UpperCAmelCase__ ) def lowercase__ ( self : Dict ): '''simple docstring''' lowercase__ = self.get_image_processor() lowercase__ = self.get_tokenizer() lowercase__ = AlignProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) lowercase__ = '''lower newer''' lowercase__ = self.prepare_image_inputs() lowercase__ = processor(text=UpperCAmelCase__, images=UpperCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), processor.model_input_names )
183
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## a__ : Optional[Any] = 1_6 a__ : str = 3_2 def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ = 16 ): '''simple docstring''' __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("bert-base-cased" ) __SCREAMING_SNAKE_CASE = load_dataset("glue" , "mrpc" ) def tokenize_function(lowerCAmelCase_ ): # max_length=None => use the model max length (it's actually the default) __SCREAMING_SNAKE_CASE = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __SCREAMING_SNAKE_CASE = datasets.map( lowerCAmelCase_ , batched=lowerCAmelCase_ , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __SCREAMING_SNAKE_CASE = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(lowerCAmelCase_ ): # On TPU it's best to pad everything to the same length or training will be very slow. __SCREAMING_SNAKE_CASE = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __SCREAMING_SNAKE_CASE = 16 elif accelerator.mixed_precision != "no": __SCREAMING_SNAKE_CASE = 8 else: __SCREAMING_SNAKE_CASE = None return tokenizer.pad( lowerCAmelCase_ , padding="longest" , max_length=lowerCAmelCase_ , pad_to_multiple_of=lowerCAmelCase_ , return_tensors="pt" , ) # Instantiate dataloaders. __SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["train"] , shuffle=lowerCAmelCase_ , collate_fn=lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["validation"] , shuffle=lowerCAmelCase_ , collate_fn=lowerCAmelCase_ , batch_size=lowerCAmelCase_ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders a__ : List[Any] = mocked_dataloaders # noqa: F811 def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' if os.environ.get("TESTING_MOCKED_DATALOADERS" , lowerCAmelCase_ ) == "1": __SCREAMING_SNAKE_CASE = 2 # Initialize accelerator __SCREAMING_SNAKE_CASE = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __SCREAMING_SNAKE_CASE = config["lr"] __SCREAMING_SNAKE_CASE = int(config["num_epochs"] ) __SCREAMING_SNAKE_CASE = int(config["seed"] ) __SCREAMING_SNAKE_CASE = int(config["batch_size"] ) __SCREAMING_SNAKE_CASE = evaluate.load("glue" , "mrpc" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=lowerCAmelCase_ ) def inner_training_loop(lowerCAmelCase_ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(lowerCAmelCase_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __SCREAMING_SNAKE_CASE = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=lowerCAmelCase_ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __SCREAMING_SNAKE_CASE = model.to(accelerator.device ) # Instantiate optimizer __SCREAMING_SNAKE_CASE = AdamW(params=model.parameters() , lr=lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = get_dataloaders(lowerCAmelCase_ , lowerCAmelCase_ ) # Instantiate scheduler __SCREAMING_SNAKE_CASE = get_linear_schedule_with_warmup( optimizer=lowerCAmelCase_ , num_warmup_steps=100 , num_training_steps=(len(lowerCAmelCase_ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = accelerator.prepare( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Now we train the model for epoch in range(lowerCAmelCase_ ): model.train() for step, batch in enumerate(lowerCAmelCase_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = outputs.loss accelerator.backward(lowerCAmelCase_ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowerCAmelCase_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = outputs.logits.argmax(dim=-1 ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=lowerCAmelCase_ , references=lowerCAmelCase_ , ) __SCREAMING_SNAKE_CASE = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , lowerCAmelCase_ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase__ (): '''simple docstring''' __SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=lowerCAmelCase_ , default=lowerCAmelCase_ , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) __SCREAMING_SNAKE_CASE = parser.parse_args() __SCREAMING_SNAKE_CASE = {"lr": 2E-5, "num_epochs": 3, "seed": 42, "batch_size": 16} training_function(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": main()
682
0
'''simple docstring''' import requests SCREAMING_SNAKE_CASE_: List[str] ='''''' # <-- Put your OpenWeatherMap appid here! SCREAMING_SNAKE_CASE_: str ='''https://api.openweathermap.org/data/2.5/''' def lowerCAmelCase_ ( snake_case_ : Optional[int] = "Chicago" , snake_case_ : Optional[Any] = APPID ) -> Dict: '''simple docstring''' return requests.get(URL_BASE + "weather" , params=locals() ).json() def lowerCAmelCase_ ( snake_case_ : Dict = "Kolkata, India" , snake_case_ : Any = APPID ) -> List[Any]: '''simple docstring''' return requests.get(URL_BASE + "forecast" , params=locals() ).json() def lowerCAmelCase_ ( snake_case_ : Optional[int] = 55.68 , snake_case_ : Dict = 12.57 , snake_case_ : Tuple = APPID ) -> List[str]: '''simple docstring''' return requests.get(URL_BASE + "onecall" , params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: SCREAMING_SNAKE_CASE_: Any =input('Enter a location:').strip() if location: pprint(current_weather(location)) else: break
78
"""simple docstring""" from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig a__ : Dict = logging.get_logger(__name__) # General docstring a__ : str = '''RegNetConfig''' # Base docstring a__ : List[str] = '''facebook/regnet-y-040''' a__ : int = [1, 1_0_8_8, 7, 7] # Image classification docstring a__ : int = '''facebook/regnet-y-040''' a__ : str = '''tabby, tabby cat''' a__ : Optional[Any] = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[str] = "relu" , **UpperCAmelCase__ : Tuple , ) -> Any: super().__init__(**UpperCAmelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb __SCREAMING_SNAKE_CASE = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) __SCREAMING_SNAKE_CASE = tf.keras.layers.ConvaD( filters=UpperCAmelCase__ , kernel_size=UpperCAmelCase__ , strides=UpperCAmelCase__ , padding="VALID" , groups=UpperCAmelCase__ , use_bias=UpperCAmelCase__ , name="convolution" , ) __SCREAMING_SNAKE_CASE = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) __SCREAMING_SNAKE_CASE = ACTaFN[activation] if activation is not None else tf.identity def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : Optional[int] ) -> Tuple: __SCREAMING_SNAKE_CASE = self.convolution(self.padding(UpperCAmelCase__ ) ) __SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase__ ) return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : List[Any] , UpperCAmelCase__ : RegNetConfig , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = config.num_channels __SCREAMING_SNAKE_CASE = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : List[Any] ) -> Optional[int]: __SCREAMING_SNAKE_CASE = shape_list(UpperCAmelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) __SCREAMING_SNAKE_CASE = tf.transpose(UpperCAmelCase__ , perm=(0, 2, 3, 1) ) __SCREAMING_SNAKE_CASE = self.embedder(UpperCAmelCase__ ) return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int = 2 , **UpperCAmelCase__ : int ) -> str: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tf.keras.layers.ConvaD( filters=UpperCAmelCase__ , kernel_size=1 , strides=UpperCAmelCase__ , use_bias=UpperCAmelCase__ , name="convolution" ) __SCREAMING_SNAKE_CASE = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : tf.Tensor , UpperCAmelCase__ : bool = False ) -> tf.Tensor: return self.normalization(self.convolution(UpperCAmelCase__ ) , training=UpperCAmelCase__ ) class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , **UpperCAmelCase__ : int ) -> Tuple: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tf.keras.layers.GlobalAveragePoolingaD(keepdims=UpperCAmelCase__ , name="pooler" ) __SCREAMING_SNAKE_CASE = [ tf.keras.layers.ConvaD(filters=UpperCAmelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=UpperCAmelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : List[str] ) -> Any: # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] __SCREAMING_SNAKE_CASE = self.pooler(UpperCAmelCase__ ) for layer_module in self.attention: __SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = hidden_state * pooled return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : Dict , UpperCAmelCase__ : RegNetConfig , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int = 1 , **UpperCAmelCase__ : int ) -> str: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = in_channels != out_channels or stride != 1 __SCREAMING_SNAKE_CASE = max(1 , out_channels // config.groups_width ) __SCREAMING_SNAKE_CASE = ( TFRegNetShortCut(UpperCAmelCase__ , stride=UpperCAmelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. __SCREAMING_SNAKE_CASE = [ TFRegNetConvLayer(UpperCAmelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( UpperCAmelCase__ , stride=UpperCAmelCase__ , groups=UpperCAmelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(UpperCAmelCase__ , kernel_size=1 , activation=UpperCAmelCase__ , name="layer.2" ), ] __SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : str ) -> Any: __SCREAMING_SNAKE_CASE = hidden_state for layer_module in self.layers: __SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.shortcut(UpperCAmelCase__ ) hidden_state += residual __SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase__ ) return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : List[str] , UpperCAmelCase__ : RegNetConfig , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int = 1 , **UpperCAmelCase__ : List[Any] ) -> Any: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = in_channels != out_channels or stride != 1 __SCREAMING_SNAKE_CASE = max(1 , out_channels // config.groups_width ) __SCREAMING_SNAKE_CASE = ( TFRegNetShortCut(UpperCAmelCase__ , stride=UpperCAmelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) __SCREAMING_SNAKE_CASE = [ TFRegNetConvLayer(UpperCAmelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( UpperCAmelCase__ , stride=UpperCAmelCase__ , groups=UpperCAmelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(UpperCAmelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(UpperCAmelCase__ , kernel_size=1 , activation=UpperCAmelCase__ , name="layer.3" ), ] __SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> List[Any]: __SCREAMING_SNAKE_CASE = hidden_state for layer_module in self.layers: __SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.shortcut(UpperCAmelCase__ ) hidden_state += residual __SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase__ ) return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : str , UpperCAmelCase__ : RegNetConfig , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int = 2 , UpperCAmelCase__ : int = 2 , **UpperCAmelCase__ : Optional[int] ) -> Optional[Any]: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer __SCREAMING_SNAKE_CASE = [ # downsampling is done in the first layer with stride of 2 layer(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , stride=UpperCAmelCase__ , name="layers.0" ), *[layer(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : int ) -> int: for layer_module in self.layers: __SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase__ ) return hidden_state class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" def __init__( self : Any , UpperCAmelCase__ : RegNetConfig , **UpperCAmelCase__ : Any ) -> List[str]: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( UpperCAmelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) __SCREAMING_SNAKE_CASE = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(UpperCAmelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , depth=UpperCAmelCase__ , name=F"""stages.{i+1}""" ) ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : tf.Tensor , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = True ) -> TFBaseModelOutputWithNoAttention: __SCREAMING_SNAKE_CASE = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __SCREAMING_SNAKE_CASE = hidden_states + (hidden_state,) __SCREAMING_SNAKE_CASE = stage_module(UpperCAmelCase__ ) if output_hidden_states: __SCREAMING_SNAKE_CASE = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=UpperCAmelCase__ , hidden_states=UpperCAmelCase__ ) @keras_serializable class UpperCamelCase_ ( tf.keras.layers.Layer): """simple docstring""" snake_case__ : Any = RegNetConfig def __init__( self : List[Any] , UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : int ) -> Tuple: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = config __SCREAMING_SNAKE_CASE = TFRegNetEmbeddings(UpperCAmelCase__ , name="embedder" ) __SCREAMING_SNAKE_CASE = TFRegNetEncoder(UpperCAmelCase__ , name="encoder" ) __SCREAMING_SNAKE_CASE = tf.keras.layers.GlobalAveragePoolingaD(keepdims=UpperCAmelCase__ , name="pooler" ) @unpack_inputs def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : tf.Tensor , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : bool = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: __SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict __SCREAMING_SNAKE_CASE = self.embedder(UpperCAmelCase__ , training=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.encoder( UpperCAmelCase__ , output_hidden_states=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , training=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = encoder_outputs[0] __SCREAMING_SNAKE_CASE = self.pooler(UpperCAmelCase__ ) # Change to NCHW output format have uniformity in the modules __SCREAMING_SNAKE_CASE = tf.transpose(UpperCAmelCase__ , perm=(0, 3, 1, 2) ) __SCREAMING_SNAKE_CASE = tf.transpose(UpperCAmelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: __SCREAMING_SNAKE_CASE = tuple([tf.transpose(UpperCAmelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase__ , pooler_output=UpperCAmelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : List[Any] = RegNetConfig snake_case__ : List[str] = "regnet" snake_case__ : str = "pixel_values" @property def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_2_4, 2_2_4) , dtype=tf.floataa )} a__ : Union[str, Any] = r''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' a__ : Optional[int] = r''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCamelCase , ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : RegNetConfig , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Optional[int] ) -> Tuple: super().__init__(UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = TFRegNetMainLayer(UpperCAmelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(UpperCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : tf.Tensor , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : Dict=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: __SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict __SCREAMING_SNAKE_CASE = self.regnet( pixel_values=UpperCAmelCase__ , output_hidden_states=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , training=UpperCAmelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCamelCase , ) class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : RegNetConfig , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> Any: super().__init__(UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = config.num_labels __SCREAMING_SNAKE_CASE = TFRegNetMainLayer(UpperCAmelCase__ , name="regnet" ) # classification head __SCREAMING_SNAKE_CASE = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(UpperCAmelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : tf.Tensor = None , UpperCAmelCase__ : tf.Tensor = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[Any]=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: __SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict __SCREAMING_SNAKE_CASE = self.regnet( UpperCAmelCase__ , output_hidden_states=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , training=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] __SCREAMING_SNAKE_CASE = self.classifier[0](UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.classifier[1](UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = None if labels is None else self.hf_compute_loss(labels=UpperCAmelCase__ , logits=UpperCAmelCase__ ) if not return_dict: __SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=UpperCAmelCase__ , logits=UpperCAmelCase__ , hidden_states=outputs.hidden_states )
682
0
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated lowerCamelCase__ = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ lowerCamelCase__ = '''https://storage.googleapis.com/cvdf-datasets/mnist/''' def UpperCamelCase ( snake_case__ : List[Any] ): '''simple docstring''' __snake_case :int = numpy.dtype(numpy.uintaa ).newbyteorder(""">""" ) return numpy.frombuffer(bytestream.read(4 ) ,dtype=lowerCAmelCase_ )[0] @deprecated(lowerCAmelCase_ ,"""Please use tf.data to implement this functionality.""" ) def UpperCamelCase ( snake_case__ : int ): '''simple docstring''' print("""Extracting""" ,f.name ) with gzip.GzipFile(fileobj=lowerCAmelCase_ ) as bytestream: __snake_case :Optional[int] = _readaa(lowerCAmelCase_ ) if magic != 2051: raise ValueError( """Invalid magic number %d in MNIST image file: %s""" % (magic, f.name) ) __snake_case :int = _readaa(lowerCAmelCase_ ) __snake_case :List[Any] = _readaa(lowerCAmelCase_ ) __snake_case :Dict = _readaa(lowerCAmelCase_ ) __snake_case :str = bytestream.read(rows * cols * num_images ) __snake_case :List[Any] = numpy.frombuffer(lowerCAmelCase_ ,dtype=numpy.uinta ) __snake_case :Union[str, Any] = data.reshape(lowerCAmelCase_ ,lowerCAmelCase_ ,lowerCAmelCase_ ,1 ) return data @deprecated(lowerCAmelCase_ ,"""Please use tf.one_hot on tensors.""" ) def UpperCamelCase ( snake_case__ : Any ,snake_case__ : List[str] ): '''simple docstring''' __snake_case :Any = labels_dense.shape[0] __snake_case :Optional[int] = numpy.arange(lowerCAmelCase_ ) * num_classes __snake_case :Any = numpy.zeros((num_labels, num_classes) ) __snake_case :Optional[int] = 1 return labels_one_hot @deprecated(lowerCAmelCase_ ,"""Please use tf.data to implement this functionality.""" ) def UpperCamelCase ( snake_case__ : int ,snake_case__ : str=False ,snake_case__ : List[Any]=10 ): '''simple docstring''' print("""Extracting""" ,f.name ) with gzip.GzipFile(fileobj=lowerCAmelCase_ ) as bytestream: __snake_case :Union[str, Any] = _readaa(lowerCAmelCase_ ) if magic != 2049: raise ValueError( """Invalid magic number %d in MNIST label file: %s""" % (magic, f.name) ) __snake_case :Any = _readaa(lowerCAmelCase_ ) __snake_case :List[str] = bytestream.read(lowerCAmelCase_ ) __snake_case :List[Any] = numpy.frombuffer(lowerCAmelCase_ ,dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(lowerCAmelCase_ ,lowerCAmelCase_ ) return labels class snake_case__ : '''simple docstring''' @deprecated( UpperCAmelCase__ , """Please use alternatives such as official/mnist/_DataSet.py""" """ from tensorflow/models.""" , ) def __init__( self , a__ , a__ , a__=False , a__=False , a__=dtypes.floataa , a__=True , a__=None , ) -> Dict: '''simple docstring''' __snake_case , __snake_case :Union[str, Any] = random_seed.get_seed(UpperCAmelCase__ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) __snake_case :Optional[int] = dtypes.as_dtype(UpperCAmelCase__ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("""Invalid image dtype %r, expected uint8 or float32""" % dtype ) if fake_data: __snake_case :Any = 1_00_00 __snake_case :str = one_hot else: assert ( images.shape[0] == labels.shape[0] ), F'''images.shape: {images.shape} labels.shape: {labels.shape}''' __snake_case :Optional[Any] = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 __snake_case :str = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. __snake_case :List[str] = images.astype(numpy.floataa ) __snake_case :str = numpy.multiply(UpperCAmelCase__ , 1.0 / 2_55.0 ) __snake_case :str = images __snake_case :List[str] = labels __snake_case :int = 0 __snake_case :Any = 0 @property def __lowercase ( self ) -> Dict: '''simple docstring''' return self._images @property def __lowercase ( self ) -> Optional[Any]: '''simple docstring''' return self._labels @property def __lowercase ( self ) -> List[Any]: '''simple docstring''' return self._num_examples @property def __lowercase ( self ) -> Tuple: '''simple docstring''' return self._epochs_completed def __lowercase ( self , a__ , a__=False , a__=True ) -> int: '''simple docstring''' if fake_data: __snake_case :List[Any] = [1] * 7_84 __snake_case :str = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(UpperCAmelCase__ )], [fake_label for _ in range(UpperCAmelCase__ )], ) __snake_case :int = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: __snake_case :List[str] = numpy.arange(self._num_examples ) numpy.random.shuffle(UpperCAmelCase__ ) __snake_case :List[Any] = self.images[perma] __snake_case :int = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch __snake_case :Union[str, Any] = self._num_examples - start __snake_case :int = self._images[start : self._num_examples] __snake_case :Tuple = self._labels[start : self._num_examples] # Shuffle the data if shuffle: __snake_case :Tuple = numpy.arange(self._num_examples ) numpy.random.shuffle(UpperCAmelCase__ ) __snake_case :Union[str, Any] = self.images[perm] __snake_case :List[str] = self.labels[perm] # Start next epoch __snake_case :Optional[int] = 0 __snake_case :int = batch_size - rest_num_examples __snake_case :Tuple = self._index_in_epoch __snake_case :int = self._images[start:end] __snake_case :Dict = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size __snake_case :Optional[Any] = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(lowerCAmelCase_ ,"""Please write your own downloading logic.""" ) def UpperCamelCase ( snake_case__ : Optional[int] ,snake_case__ : Any ,snake_case__ : Any ): '''simple docstring''' if not gfile.Exists(lowerCAmelCase_ ): gfile.MakeDirs(lowerCAmelCase_ ) __snake_case :Optional[Any] = os.path.join(lowerCAmelCase_ ,lowerCAmelCase_ ) if not gfile.Exists(lowerCAmelCase_ ): urllib.request.urlretrieve(lowerCAmelCase_ ,lowerCAmelCase_ ) # noqa: S310 with gfile.GFile(lowerCAmelCase_ ) as f: __snake_case :Tuple = f.size() print("""Successfully downloaded""" ,lowerCAmelCase_ ,lowerCAmelCase_ ,"""bytes.""" ) return filepath @deprecated( lowerCAmelCase_ ,"""Please use alternatives such as:""" """ tensorflow_datasets.load('mnist')""" ) def UpperCamelCase ( snake_case__ : Optional[int] ,snake_case__ : Any=False ,snake_case__ : Tuple=False ,snake_case__ : Any=dtypes.floataa ,snake_case__ : Dict=True ,snake_case__ : str=5000 ,snake_case__ : Any=None ,snake_case__ : Dict=DEFAULT_SOURCE_URL ,): '''simple docstring''' if fake_data: def fake(): return _DataSet( [] ,[] ,fake_data=lowerCAmelCase_ ,one_hot=lowerCAmelCase_ ,dtype=lowerCAmelCase_ ,seed=lowerCAmelCase_ ) __snake_case :Tuple = fake() __snake_case :Union[str, Any] = fake() __snake_case :List[str] = fake() return _Datasets(train=lowerCAmelCase_ ,validation=lowerCAmelCase_ ,test=lowerCAmelCase_ ) if not source_url: # empty string check __snake_case :Union[str, Any] = DEFAULT_SOURCE_URL __snake_case :List[Any] = """train-images-idx3-ubyte.gz""" __snake_case :Optional[int] = """train-labels-idx1-ubyte.gz""" __snake_case :str = """t10k-images-idx3-ubyte.gz""" __snake_case :Optional[Any] = """t10k-labels-idx1-ubyte.gz""" __snake_case :List[Any] = _maybe_download( lowerCAmelCase_ ,lowerCAmelCase_ ,source_url + train_images_file ) with gfile.Open(lowerCAmelCase_ ,"""rb""" ) as f: __snake_case :str = _extract_images(lowerCAmelCase_ ) __snake_case :Dict = _maybe_download( lowerCAmelCase_ ,lowerCAmelCase_ ,source_url + train_labels_file ) with gfile.Open(lowerCAmelCase_ ,"""rb""" ) as f: __snake_case :Optional[Any] = _extract_labels(lowerCAmelCase_ ,one_hot=lowerCAmelCase_ ) __snake_case :List[str] = _maybe_download( lowerCAmelCase_ ,lowerCAmelCase_ ,source_url + test_images_file ) with gfile.Open(lowerCAmelCase_ ,"""rb""" ) as f: __snake_case :List[str] = _extract_images(lowerCAmelCase_ ) __snake_case :List[Any] = _maybe_download( lowerCAmelCase_ ,lowerCAmelCase_ ,source_url + test_labels_file ) with gfile.Open(lowerCAmelCase_ ,"""rb""" ) as f: __snake_case :Tuple = _extract_labels(lowerCAmelCase_ ,one_hot=lowerCAmelCase_ ) if not 0 <= validation_size <= len(lowerCAmelCase_ ): __snake_case :Union[str, Any] = ( """Validation size should be between 0 and """ f'''{len(lowerCAmelCase_ )}. Received: {validation_size}.''' ) raise ValueError(lowerCAmelCase_ ) __snake_case :Optional[Any] = train_images[:validation_size] __snake_case :Any = train_labels[:validation_size] __snake_case :List[Any] = train_images[validation_size:] __snake_case :Optional[int] = train_labels[validation_size:] __snake_case :Any = {"""dtype""": dtype, """reshape""": reshape, """seed""": seed} __snake_case :Optional[int] = _DataSet(lowerCAmelCase_ ,lowerCAmelCase_ ,**lowerCAmelCase_ ) __snake_case :Tuple = _DataSet(lowerCAmelCase_ ,lowerCAmelCase_ ,**lowerCAmelCase_ ) __snake_case :Tuple = _DataSet(lowerCAmelCase_ ,lowerCAmelCase_ ,**lowerCAmelCase_ ) return _Datasets(train=lowerCAmelCase_ ,validation=lowerCAmelCase_ ,test=lowerCAmelCase_ )
455
"""simple docstring""" def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError("Length must be a positive." ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError("Length must be a positive." ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
682
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowercase_ (unittest.TestCase ): @slow def __UpperCamelCase ( self) -> str: a__ =AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=UpperCAmelCase__).to(UpperCAmelCase__) a__ =AutoTokenizer.from_pretrained('google/mt5-small') a__ =tokenizer('Hello there' , return_tensors='pt').input_ids a__ =tokenizer('Hi I am' , return_tensors='pt').input_ids a__ =model(input_ids.to(UpperCAmelCase__) , labels=labels.to(UpperCAmelCase__)).loss a__ =-(labels.shape[-1] * loss.item()) a__ =-84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
20
"""simple docstring""" import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging a__ : Union[str, Any] = logging.get_logger(__name__) a__ : Optional[int] = r''' Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. ''' class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" @add_start_docstrings(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : torch.LongTensor , UpperCAmelCase__ : torch.FloatTensor , **UpperCAmelCase__ : List[Any] ) -> bool: raise NotImplementedError("StoppingCriteria needs to be subclassed" ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] = None ) -> Optional[int]: __SCREAMING_SNAKE_CASE = max_length __SCREAMING_SNAKE_CASE = max_position_embeddings @add_start_docstrings(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : torch.LongTensor , UpperCAmelCase__ : torch.FloatTensor , **UpperCAmelCase__ : Optional[int] ) -> bool: __SCREAMING_SNAKE_CASE = input_ids.shape[-1] __SCREAMING_SNAKE_CASE = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( "This is a friendly reminder - the current text generation call will exceed the model's predefined " F"""maximum length ({self.max_position_embeddings}). Depending on the model, you may observe """ "exceptions, performance degradation, or nothing at all." ) return is_done class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> str: warnings.warn( "The class `MaxNewTokensCriteria` is deprecated. " F"""Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` """ "with `max_length = start_length + max_new_tokens` instead." , UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = start_length __SCREAMING_SNAKE_CASE = max_new_tokens __SCREAMING_SNAKE_CASE = start_length + max_new_tokens @add_start_docstrings(UpperCAmelCase__ ) def __call__( self : Union[str, Any] , UpperCAmelCase__ : torch.LongTensor , UpperCAmelCase__ : torch.FloatTensor , **UpperCAmelCase__ : Tuple ) -> bool: return input_ids.shape[-1] >= self.max_length class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : float , UpperCAmelCase__ : Optional[float] = None ) -> Dict: __SCREAMING_SNAKE_CASE = max_time __SCREAMING_SNAKE_CASE = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(UpperCAmelCase__ ) def __call__( self : Tuple , UpperCAmelCase__ : torch.LongTensor , UpperCAmelCase__ : torch.FloatTensor , **UpperCAmelCase__ : str ) -> bool: return time.time() - self.initial_timestamp > self.max_time class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" @add_start_docstrings(UpperCAmelCase__ ) def __call__( self : Dict , UpperCAmelCase__ : torch.LongTensor , UpperCAmelCase__ : torch.FloatTensor , **UpperCAmelCase__ : List[str] ) -> bool: return any(criteria(UpperCAmelCase__ , UpperCAmelCase__ ) for criteria in self ) @property def UpperCAmelCase_ ( self : Any ) -> Optional[int]: for stopping_criterium in self: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): return stopping_criterium.max_length elif isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): return stopping_criterium.max_length return None def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = stopping_criteria.max_length __SCREAMING_SNAKE_CASE = deepcopy(lowerCAmelCase_ ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn("You set different `max_length` for stopping criteria and `max_length` parameter" , lowerCAmelCase_ ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=lowerCAmelCase_ ) ) return new_stopping_criteria
682
0
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' def __init__( self : int , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Optional[Any] ): _a = params _a = np.array(UpperCAmelCase__ ) _a = np.array([len(UpperCAmelCase__ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : List[str] , SCREAMING_SNAKE_CASE_ : str ): return (self.token_ids[index], self.lengths[index]) def __len__( self : Tuple ): return len(self.lengths ) def _UpperCAmelCase ( self : Dict ): assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _UpperCAmelCase ( self : Union[str, Any] ): _a = self.params.max_model_input_size _a = self.lengths > max_len logger.info(f"""Splitting {sum(UpperCAmelCase__ )} too long sequences.""" ) def divide_chunks(SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any] ): return [l[i : i + n] for i in range(0 , len(UpperCAmelCase__ ) , UpperCAmelCase__ )] _a = [] _a = [] if self.params.mlm: _a , _a = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: _a , _a = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: _a = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: _a = np.insert(UpperCAmelCase__ , 0 , UpperCAmelCase__ ) if sub_s[-1] != sep_id: _a = np.insert(UpperCAmelCase__ , len(UpperCAmelCase__ ) , UpperCAmelCase__ ) assert len(UpperCAmelCase__ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(UpperCAmelCase__ ) new_tok_ids.extend(UpperCAmelCase__ ) new_lengths.extend([len(UpperCAmelCase__ ) for l in sub_seqs] ) _a = np.array(UpperCAmelCase__ ) _a = np.array(UpperCAmelCase__ ) def _UpperCAmelCase ( self : List[str] ): _a = len(self ) _a = self.lengths > 1_1 _a = self.token_ids[indices] _a = self.lengths[indices] _a = len(self ) logger.info(f"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def _UpperCAmelCase ( self : Union[str, Any] ): if "unk_token" not in self.params.special_tok_ids: return else: _a = self.params.special_tok_ids['unk_token'] _a = len(self ) _a = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) _a = (unk_occs / self.lengths) < 0.5 _a = self.token_ids[indices] _a = self.lengths[indices] _a = len(self ) logger.info(f"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def _UpperCAmelCase ( self : List[Any] ): if not self.params.is_master: return logger.info(f"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _UpperCAmelCase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[str] ): _a = [t[0] for t in batch] _a = [t[1] for t in batch] assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) # Max for paddings _a = max(UpperCAmelCase__ ) # Pad token ids if self.params.mlm: _a = self.params.special_tok_ids['pad_token'] else: _a = self.params.special_tok_ids['unk_token'] _a = [list(t.astype(UpperCAmelCase__ ) ) + [pad_idx] * (max_seq_len_ - len(UpperCAmelCase__ )) for t in token_ids] assert len(tk_ ) == len(UpperCAmelCase__ ) assert all(len(UpperCAmelCase__ ) == max_seq_len_ for t in tk_ ) _a = torch.tensor(tk_ ) # (bs, max_seq_len_) _a = torch.tensor(UpperCAmelCase__ ) # (bs) return tk_t, lg_t
562
"""simple docstring""" import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : int = RoCBertTokenizer snake_case__ : int = None snake_case__ : Optional[Any] = False snake_case__ : int = True snake_case__ : Any = filter_non_english def UpperCAmelCase_ ( self : Any ) -> Union[str, Any]: super().setUp() __SCREAMING_SNAKE_CASE = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "你", "好", "是", "谁", "a", "b", "c", "d"] __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = {} for i, value in enumerate(UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = i __SCREAMING_SNAKE_CASE = i __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_shape_file"] ) __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_pronunciation_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) with open(self.word_shape_file , "w" , encoding="utf-8" ) as word_shape_writer: json.dump(UpperCAmelCase__ , UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) with open(self.word_pronunciation_file , "w" , encoding="utf-8" ) as word_pronunciation_writer: json.dump(UpperCAmelCase__ , UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> List[str]: __SCREAMING_SNAKE_CASE = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("你好[SEP]你是谁" ) self.assertListEqual(UpperCAmelCase__ , ["你", "好", "[SEP]", "你", "是", "谁"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(UpperCAmelCase__ ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(UpperCAmelCase__ ) , [5, 6, 2, 5, 7, 8] ) def UpperCAmelCase_ ( self : Any ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase_ ( self : Dict ) -> Dict: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def UpperCAmelCase_ ( self : int ) -> Dict: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase_ ( self : Any ) -> Optional[int]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase_ ( self : int ) -> List[Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = RoCBertBasicTokenizer(do_lower_case=UpperCAmelCase__ , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def UpperCAmelCase_ ( self : str ) -> List[str]: __SCREAMING_SNAKE_CASE = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] __SCREAMING_SNAKE_CASE = {} for i, token in enumerate(UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = i __SCREAMING_SNAKE_CASE = RoCBertWordpieceTokenizer(vocab=UpperCAmelCase__ , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def UpperCAmelCase_ ( self : List[Any] ) -> str: self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def UpperCAmelCase_ ( self : List[str] ) -> Tuple: self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def UpperCAmelCase_ ( self : int ) -> int: __SCREAMING_SNAKE_CASE = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(UpperCAmelCase__ ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) if self.test_rust_tokenizer: __SCREAMING_SNAKE_CASE = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(UpperCAmelCase__ ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __SCREAMING_SNAKE_CASE = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = F"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" __SCREAMING_SNAKE_CASE = tokenizer_r.encode_plus( UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_offsets_mapping=UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , ) __SCREAMING_SNAKE_CASE = tokenizer_r.do_lower_case if hasattr(UpperCAmelCase__ , "do_lower_case" ) else False __SCREAMING_SNAKE_CASE = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), "Allen"), ((2_1, 2_3), "##NL"), ((2_3, 2_4), "##P"), ((2_5, 3_3), "sentence"), ((3_3, 3_4), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), "allen"), ((2_1, 2_3), "##nl"), ((2_3, 2_4), "##p"), ((2_5, 3_3), "sentence"), ((3_3, 3_4), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def UpperCAmelCase_ ( self : Tuple ) -> Dict: __SCREAMING_SNAKE_CASE = ["的", "人", "有"] __SCREAMING_SNAKE_CASE = "".join(UpperCAmelCase__ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_p.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_r.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_r.convert_ids_to_tokens(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_p.convert_ids_to_tokens(UpperCAmelCase__ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_r.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_p.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_r.convert_ids_to_tokens(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer_p.convert_ids_to_tokens(UpperCAmelCase__ ) # it is expected that only the first Chinese character is not preceded by "##". __SCREAMING_SNAKE_CASE = [ F"""##{token}""" if idx != 0 else token for idx, token in enumerate(UpperCAmelCase__ ) ] self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def UpperCAmelCase_ ( self : List[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __SCREAMING_SNAKE_CASE = tokenizer.encode("你好" , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.encode("你是谁" , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ , UpperCAmelCase__ ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def UpperCAmelCase_ ( self : str ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.get_tokenizers(do_lower_case=UpperCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): __SCREAMING_SNAKE_CASE = "你好,你是谁" __SCREAMING_SNAKE_CASE = tokenizer.tokenize(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_shape_ids(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_pronunciation_ids(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.prepare_for_model( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.encode_plus(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
682
0
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def A ( lowercase__ : int , lowercase__ : Tuple , lowercase__ : Optional[int] , lowercase__ : Any , lowercase__ : Union[str, Any] , lowercase__ : int ) -> str: if (ksize % 2) == 0: UpperCamelCase__ :Union[str, Any] = ksize + 1 UpperCamelCase__ :Optional[int] = np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(lowerCAmelCase_ ): for x in range(lowerCAmelCase_ ): # distance from center UpperCamelCase__ :str = x - ksize // 2 UpperCamelCase__ :Dict = y - ksize // 2 # degree to radiant UpperCamelCase__ :int = theta / 180 * np.pi UpperCamelCase__ :Any = np.cos(_theta ) UpperCamelCase__ :int = np.sin(_theta ) # get kernel x UpperCamelCase__ :Dict = cos_theta * px + sin_theta * py # get kernel y UpperCamelCase__ :Any = -sin_theta * px + cos_theta * py # fill kernel UpperCamelCase__ :int = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image UpperCamelCase = imread("../image_data/lena.jpg") # turn image in gray scale value UpperCamelCase = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges UpperCamelCase = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 120, 150]: UpperCamelCase = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) UpperCamelCase = out / out.max() * 255 UpperCamelCase = out.astype(np.uinta) imshow("Original", gray) imshow("Gabor filter with 20x20 mask and 6 directions", out) waitKey(0)
45
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Union[str, Any] = logging.get_logger(__name__) a__ : Optional[int] = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Optional[int] = "vivit" def __init__( self : Dict , UpperCAmelCase__ : Dict=2_2_4 , UpperCAmelCase__ : List[Any]=3_2 , UpperCAmelCase__ : str=[2, 1_6, 1_6] , UpperCAmelCase__ : str=3 , UpperCAmelCase__ : str=7_6_8 , UpperCAmelCase__ : Dict=1_2 , UpperCAmelCase__ : Optional[int]=1_2 , UpperCAmelCase__ : Any=3_0_7_2 , UpperCAmelCase__ : Optional[int]="gelu_fast" , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Union[str, Any]=0.0 , UpperCAmelCase__ : Optional[int]=0.02 , UpperCAmelCase__ : str=1E-06 , UpperCAmelCase__ : List[Any]=True , **UpperCAmelCase__ : Any , ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = num_frames __SCREAMING_SNAKE_CASE = tubelet_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = qkv_bias super().__init__(**UpperCAmelCase__ )
682
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue_model_parallelism.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "roberta-large", "instance_type": "ml.p3dn.24xlarge", "results": {"train_runtime": 1_600, "eval_accuracy": 0.3, "eval_loss": 1.2}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self : Any ) -> List[Any]: if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=UpperCAmelCase__ , ) assert hasattr(self , "env" ) def __SCREAMING_SNAKE_CASE ( self : str , __a : List[str] ) -> List[Any]: # configuration for running training on smdistributed Model Parallel _UpperCamelCase : List[str] = { "enabled": True, "processes_per_host": 8, } _UpperCamelCase : int = { "enabled": True, "parameters": { "microbatches": 4, "placement_strategy": "spread", "pipeline": "interleaved", "optimize": "speed", "partitions": 4, "ddp": True, }, } _UpperCamelCase : Any = {"smdistributed": {"modelparallel": smp_options}, "mpi": mpi_options} _UpperCamelCase : List[Any] = "trainer" if self.script == "run_glue.py" else "smtrainer" # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-{instance_count}-smp-{name_extension}''' , instance_count=UpperCAmelCase__ , instance_type=self.instance_type , debugger_hook_config=UpperCAmelCase__ , hyperparameters={ **self.env.hyperparameters, "model_name_or_path": self.model_name_or_path, "max_steps": 500, } , metric_definitions=self.env.metric_definitions , distribution=UpperCAmelCase__ , py_version="py36" , ) def __SCREAMING_SNAKE_CASE ( self : List[Any] , __a : List[str] ) -> Tuple: TrainingJobAnalytics(UpperCAmelCase__ ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(1,)] ) def __SCREAMING_SNAKE_CASE ( self : List[str] , __a : List[Any] ) -> Union[str, Any]: # create estimator _UpperCamelCase : Optional[Any] = self.create_estimator(UpperCAmelCase__ ) # run training estimator.fit() # result dataframe _UpperCamelCase : Optional[int] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _UpperCamelCase : Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _UpperCamelCase : Any = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _UpperCamelCase : str = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , UpperCAmelCase__ )
624
"""simple docstring""" import numpy as np from transformers import Pipeline def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = np.max(lowerCAmelCase_ , axis=-1 , keepdims=lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = np.exp(outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=lowerCAmelCase_ ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def UpperCAmelCase_ ( self : Tuple , **UpperCAmelCase__ : str ) -> Optional[int]: __SCREAMING_SNAKE_CASE = {} if "second_text" in kwargs: __SCREAMING_SNAKE_CASE = kwargs["second_text"] return preprocess_kwargs, {}, {} def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict=None ) -> str: return self.tokenizer(UpperCAmelCase__ , text_pair=UpperCAmelCase__ , return_tensors=self.framework ) def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Optional[Any] ) -> List[Any]: return self.model(**UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: __SCREAMING_SNAKE_CASE = model_outputs.logits[0].numpy() __SCREAMING_SNAKE_CASE = softmax(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = np.argmax(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.model.config.idalabel[best_class] __SCREAMING_SNAKE_CASE = probabilities[best_class].item() __SCREAMING_SNAKE_CASE = logits.tolist() return {"label": label, "score": score, "logits": logits}
682
0
"""simple docstring""" import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCamelCase : '''simple docstring''' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=1_28 , __a=32 , __a=16 , __a=2 , __a=0.0_2 , __a=3 , __a=4 , __a=None , ): __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_labels __lowerCAmelCase = num_choices __lowerCAmelCase = scope def snake_case ( self ): __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self ): return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , ) def snake_case ( self ): ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = self.prepare_config_and_inputs() __lowerCAmelCase = True __lowerCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = NezhaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) __lowerCAmelCase = model(UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) __lowerCAmelCase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a , __a , ): __lowerCAmelCase = True __lowerCAmelCase = NezhaModel(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , encoder_attention_mask=UpperCAmelCase__ , ) __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , ) __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = NezhaForMaskedLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = NezhaForNextSentencePrediction(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = NezhaForPreTraining(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , next_sentence_label=UpperCAmelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = NezhaForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = NezhaForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = NezhaForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_choices __lowerCAmelCase = NezhaForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowerCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case ( self ): __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : str =( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __UpperCAmelCase : Tuple =( { "feature-extraction": NezhaModel, "fill-mask": NezhaForMaskedLM, "question-answering": NezhaForQuestionAnswering, "text-classification": NezhaForSequenceClassification, "token-classification": NezhaForTokenClassification, "zero-shot": NezhaForSequenceClassification, } if is_torch_available() else {} ) __UpperCAmelCase : int =True def snake_case ( self , __a , __a , __a=False ): __lowerCAmelCase = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) if return_labels: if model_class in get_values(UpperCAmelCase__ ): __lowerCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=UpperCAmelCase__ ) __lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase__ ) return inputs_dict def snake_case ( self ): __lowerCAmelCase = NezhaModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37 ) def snake_case ( self ): self.config_tester.run_common_tests() def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*UpperCAmelCase__ ) def snake_case ( self ): # This regression test was failing with PyTorch < 1.3 ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() __lowerCAmelCase = None self.model_tester.create_and_check_model_as_decoder( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase__ ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase__ ) @slow def snake_case ( self ): for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = NezhaModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @slow @require_torch_gpu def snake_case ( self ): __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return __lowerCAmelCase = True __lowerCAmelCase = model_class(config=UpperCAmelCase__ ) __lowerCAmelCase = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) __lowerCAmelCase = torch.jit.trace( UpperCAmelCase__ , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(UpperCAmelCase__ , os.path.join(UpperCAmelCase__ , "bert.pt" ) ) __lowerCAmelCase = torch.jit.load(os.path.join(UpperCAmelCase__ , "bert.pt" ) , map_location=UpperCAmelCase__ ) loaded(inputs_dict["input_ids"].to(UpperCAmelCase__ ) , inputs_dict["attention_mask"].to(UpperCAmelCase__ ) ) @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def snake_case ( self ): __lowerCAmelCase = NezhaModel.from_pretrained("sijunhe/nezha-cn-base" ) __lowerCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __lowerCAmelCase = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] __lowerCAmelCase = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __lowerCAmelCase = torch.tensor([[[0.0_6_8_5, 0.2_4_4_1, 0.1_1_0_2], [0.0_6_0_0, 0.1_9_0_6, 0.1_3_4_9], [0.0_2_2_1, 0.0_8_1_9, 0.0_5_8_6]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCAmelCase__ , atol=1e-4 ) ) @slow def snake_case ( self ): __lowerCAmelCase = NezhaForMaskedLM.from_pretrained("sijunhe/nezha-cn-base" ) __lowerCAmelCase = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __lowerCAmelCase = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] __lowerCAmelCase = torch.Size((1, 6, 2_11_28) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __lowerCAmelCase = torch.tensor( [[-2.7_9_3_9, -1.7_9_0_2, -2.2_1_8_9], [-2.8_5_8_5, -1.8_9_0_8, -2.3_7_2_3], [-2.6_4_9_9, -1.7_7_5_0, -2.2_5_5_8]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCAmelCase__ , atol=1e-4 ) )
636
"""simple docstring""" import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('''.''') def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( "`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got " f"""{test_file} instead.""" ) __SCREAMING_SNAKE_CASE = components[-1] if not test_fn.endswith("py" ): raise ValueError(f"""`test_file` should be a python file. Got {test_fn} instead.""" ) if not test_fn.startswith("test_modeling_" ): raise ValueError( f"""`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.""" ) __SCREAMING_SNAKE_CASE = components[:-1] + [test_fn.replace(".py" , "" )] __SCREAMING_SNAKE_CASE = ".".join(lowerCAmelCase_ ) return test_module_path def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_module_path(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = importlib.import_module(lowerCAmelCase_ ) return test_module def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = get_test_module(lowerCAmelCase_ ) for attr in dir(lowerCAmelCase_ ): if attr.endswith("ModelTester" ): tester_classes.append(getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) # sort with class names return sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x.__name__ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = get_test_module(lowerCAmelCase_ ) for attr in dir(lowerCAmelCase_ ): __SCREAMING_SNAKE_CASE = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). __SCREAMING_SNAKE_CASE = getattr(lowerCAmelCase_ , "all_model_classes" , [] ) if len(lowerCAmelCase_ ) > 0: test_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x.__name__ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_test_classes(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x.__name__ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = test_class() if hasattr(lowerCAmelCase_ , "setUp" ): test.setUp() __SCREAMING_SNAKE_CASE = None if hasattr(lowerCAmelCase_ , "model_tester" ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: __SCREAMING_SNAKE_CASE = test.model_tester.__class__ return model_tester def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_test_classes(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x.__name__ ) def UpperCAmelCase__ (lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_test_classes_for_model(lowerCAmelCase_ , lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = [] for test_class in test_classes: __SCREAMING_SNAKE_CASE = get_model_tester_from_test_class(lowerCAmelCase_ ) if tester_class is not None: tester_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x.__name__ ) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_test_classes(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = {test_class: get_model_tester_from_test_class(lowerCAmelCase_ ) for test_class in test_classes} return test_tester_mapping def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_model_classes(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = { model_class: get_test_classes_for_model(lowerCAmelCase_ , lowerCAmelCase_ ) for model_class in model_classes } return model_test_mapping def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_model_classes(lowerCAmelCase_ ) __SCREAMING_SNAKE_CASE = { model_class: get_tester_classes_for_model(lowerCAmelCase_ , lowerCAmelCase_ ) for model_class in model_classes } return model_to_tester_mapping def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return o elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return o.__name__ elif isinstance(lowerCAmelCase_ , (list, tuple) ): return [to_json(lowerCAmelCase_ ) for x in o] elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return {to_json(lowerCAmelCase_ ): to_json(lowerCAmelCase_ ) for k, v in o.items()} else: return o
682
0
'''simple docstring''' import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def snake_case__ ( UpperCamelCase ) -> Optional[Any]: if isinstance(UpperCamelCase ,collections.abc.Iterable ): return x return (x, x) @require_flax class UpperCAmelCase : """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> str: pass def _lowercase ( self ) -> Optional[int]: pass def _lowercase ( self ) -> Dict: pass def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Dict: _UpperCamelCase : int = np.abs((a - b) ).max() self.assertLessEqual(_snake_case , _snake_case , F'''Difference between torch and flax is {diff} (>= {tol}).''' ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=None , **_snake_case ) -> Union[str, Any]: _UpperCamelCase : str = VisionTextDualEncoderConfig.from_vision_text_configs(_snake_case , _snake_case ) _UpperCamelCase : List[str] = FlaxVisionTextDualEncoderModel(_snake_case ) _UpperCamelCase : Union[str, Any] = model(input_ids=_snake_case , pixel_values=_snake_case , attention_mask=_snake_case ) self.assertEqual(output['''text_embeds'''].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape , (pixel_values.shape[0], config.projection_dim) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=None , **_snake_case ) -> Dict: _UpperCamelCase, _UpperCamelCase : Any = self.get_vision_text_model(_snake_case , _snake_case ) _UpperCamelCase : int = {'''vision_model''': vision_model, '''text_model''': text_model} _UpperCamelCase : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_snake_case ) _UpperCamelCase : Optional[int] = model(input_ids=_snake_case , pixel_values=_snake_case , attention_mask=_snake_case ) self.assertEqual(output['''text_embeds'''].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['''image_embeds'''].shape , (pixel_values.shape[0], model.config.projection_dim) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=None , **_snake_case ) -> int: _UpperCamelCase, _UpperCamelCase : Any = self.get_vision_text_model(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = {'''vision_model''': vision_model, '''text_model''': text_model} _UpperCamelCase : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_snake_case ) _UpperCamelCase : Union[str, Any] = model(input_ids=_snake_case , pixel_values=_snake_case , attention_mask=_snake_case ) _UpperCamelCase : Any = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_snake_case ) _UpperCamelCase : List[str] = FlaxVisionTextDualEncoderModel.from_pretrained(_snake_case ) _UpperCamelCase : Optional[int] = model(input_ids=_snake_case , pixel_values=_snake_case , attention_mask=_snake_case ) _UpperCamelCase : str = after_output[0] _UpperCamelCase : str = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_snake_case , 1E-3 ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=None , **_snake_case ) -> str: _UpperCamelCase, _UpperCamelCase : str = self.get_vision_text_model(_snake_case , _snake_case ) _UpperCamelCase : List[Any] = {'''vision_model''': vision_model, '''text_model''': text_model} _UpperCamelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**_snake_case ) _UpperCamelCase : Any = model( input_ids=_snake_case , pixel_values=_snake_case , attention_mask=_snake_case , output_attentions=_snake_case ) _UpperCamelCase : Optional[int] = output.vision_model_output.attentions self.assertEqual(len(_snake_case ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCamelCase : Optional[Any] = to_atuple(vision_model.config.image_size ) _UpperCamelCase : Dict = to_atuple(vision_model.config.patch_size ) _UpperCamelCase : Tuple = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _UpperCamelCase : Any = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _UpperCamelCase : str = output.text_model_output.attentions self.assertEqual(len(_snake_case ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Optional[int]: pt_model.to(_snake_case ) pt_model.eval() # prepare inputs _UpperCamelCase : List[Any] = inputs_dict _UpperCamelCase : Union[str, Any] = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): _UpperCamelCase : Optional[Any] = pt_model(**_snake_case ).to_tuple() _UpperCamelCase : Optional[int] = fx_model(**_snake_case ).to_tuple() self.assertEqual(len(_snake_case ) , len(_snake_case ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(_snake_case , pt_output.numpy() , 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_snake_case ) _UpperCamelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_snake_case , from_pt=_snake_case ) _UpperCamelCase : Optional[int] = fx_model_loaded(**_snake_case ).to_tuple() self.assertEqual(len(_snake_case ) , len(_snake_case ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(_snake_case , pt_output.numpy() , 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_snake_case ) _UpperCamelCase : Tuple = VisionTextDualEncoderModel.from_pretrained(_snake_case , from_flax=_snake_case ) pt_model_loaded.to(_snake_case ) pt_model_loaded.eval() with torch.no_grad(): _UpperCamelCase : str = pt_model_loaded(**_snake_case ).to_tuple() self.assertEqual(len(_snake_case ) , len(_snake_case ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(_snake_case , pt_output_loaded.numpy() , 4E-2 ) def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> str: _UpperCamelCase : List[str] = VisionTextDualEncoderConfig.from_vision_text_configs(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = VisionTextDualEncoderModel(_snake_case ) _UpperCamelCase : str = FlaxVisionTextDualEncoderModel(_snake_case ) _UpperCamelCase : Optional[int] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _snake_case ) _UpperCamelCase : List[str] = fx_state self.check_pt_flax_equivalence(_snake_case , _snake_case , _snake_case ) def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Tuple: _UpperCamelCase : List[Any] = VisionTextDualEncoderConfig.from_vision_text_configs(_snake_case , _snake_case ) _UpperCamelCase : Tuple = VisionTextDualEncoderModel(_snake_case ) _UpperCamelCase : Optional[Any] = FlaxVisionTextDualEncoderModel(_snake_case ) _UpperCamelCase : int = load_flax_weights_in_pytorch_model(_snake_case , fx_model.params ) self.check_pt_flax_equivalence(_snake_case , _snake_case , _snake_case ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_snake_case ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Optional[Any] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_snake_case ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**_snake_case ) def _lowercase ( self ) -> Dict: _UpperCamelCase : str = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_snake_case ) @is_pt_flax_cross_test def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() _UpperCamelCase : Optional[int] = config_inputs_dict.pop('''vision_config''' ) _UpperCamelCase : int = config_inputs_dict.pop('''text_config''' ) _UpperCamelCase : List[Any] = config_inputs_dict self.check_equivalence_pt_to_flax(_snake_case , _snake_case , _snake_case ) self.check_equivalence_flax_to_pt(_snake_case , _snake_case , _snake_case ) @slow def _lowercase ( self ) -> Optional[Any]: _UpperCamelCase, _UpperCamelCase : str = self.get_pretrained_model_and_inputs() _UpperCamelCase : List[str] = model_a(**_snake_case ) _UpperCamelCase : List[Any] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_snake_case ) _UpperCamelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_pretrained(_snake_case ) _UpperCamelCase : Union[str, Any] = model_a(**_snake_case ) _UpperCamelCase : List[Any] = after_outputs[0] _UpperCamelCase : List[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_snake_case , 1E-5 ) @require_flax class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: _UpperCamelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-vit''' , '''hf-internal-testing/tiny-bert''' , vision_from_pt=_snake_case , text_from_pt=_snake_case , ) _UpperCamelCase : Dict = 13 _UpperCamelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _UpperCamelCase : Any = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _UpperCamelCase : str = random_attention_mask([batch_size, 4] ) _UpperCamelCase : Optional[Any] = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def _lowercase ( self , _snake_case , _snake_case ) -> Any: _UpperCamelCase : List[Any] = FlaxViTModel(_snake_case ) _UpperCamelCase : Optional[int] = FlaxBertModel(_snake_case ) return vision_model, text_model def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = FlaxViTModelTester(self ) _UpperCamelCase : int = FlaxBertModelTester(self ) _UpperCamelCase : int = vit_model_tester.prepare_config_and_inputs() _UpperCamelCase : str = bert_model_tester.prepare_config_and_inputs() _UpperCamelCase, _UpperCamelCase : str = vision_config_and_inputs _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Union[str, Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-clip''' , '''hf-internal-testing/tiny-bert''' , vision_from_pt=_snake_case , text_from_pt=_snake_case , ) _UpperCamelCase : Tuple = 13 _UpperCamelCase : Optional[Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _UpperCamelCase : Tuple = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _UpperCamelCase : Dict = random_attention_mask([batch_size, 4] ) _UpperCamelCase : Dict = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def _lowercase ( self , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : Dict = FlaxCLIPVisionModel(_snake_case ) _UpperCamelCase : List[str] = FlaxBertModel(_snake_case ) return vision_model, text_model def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Dict = FlaxCLIPVisionModelTester(self ) _UpperCamelCase : str = FlaxBertModelTester(self ) _UpperCamelCase : Tuple = clip_model_tester.prepare_config_and_inputs() _UpperCamelCase : Tuple = bert_model_tester.prepare_config_and_inputs() _UpperCamelCase, _UpperCamelCase : int = vision_config_and_inputs _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[Any] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self ) -> int: _UpperCamelCase : str = FlaxVisionTextDualEncoderModel.from_pretrained('''clip-italian/clip-italian''' , logit_scale_init_value=1.0 ) _UpperCamelCase : List[Any] = VisionTextDualEncoderProcessor.from_pretrained('''clip-italian/clip-italian''' ) _UpperCamelCase : Optional[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) _UpperCamelCase : Optional[int] = processor( text=['''una foto di un gatto''', '''una foto di un cane'''] , images=_snake_case , padding=_snake_case , return_tensors='''np''' ) _UpperCamelCase : int = model(**_snake_case ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _UpperCamelCase : List[Any] = np.array([[1.2_284_727, 0.3_104_122]] ) self.assertTrue(np.allclose(outputs.logits_per_image , _snake_case , atol=1E-3 ) )
683
'''simple docstring''' _UpperCAmelCase : str = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : List[str] = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> str: assert len(str(UpperCamelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _UpperCamelCase : Any = year // 1_00 _UpperCamelCase : List[Any] = (5 * (century % 4) + 2) % 7 _UpperCamelCase : Tuple = year % 1_00 _UpperCamelCase : Optional[int] = centurian % 12 _UpperCamelCase : Tuple = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _UpperCamelCase : List[Any] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) _UpperCamelCase : Optional[int] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: if height >= 1: move_tower(height - 1 ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) move_disk(UpperCamelCase ,UpperCamelCase ) move_tower(height - 1 ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: print('''moving disk from''' ,UpperCamelCase ,'''to''' ,UpperCamelCase ) def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : Union[str, Any] = int(input('''Height of hanoi: ''' ).strip() ) move_tower(UpperCamelCase ,'''A''' ,'''B''' ,'''C''' ) if __name__ == "__main__": main()
683
'''simple docstring''' import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : """simple docstring""" @staticmethod def _lowercase ( *_snake_case , **_snake_case ) -> str: pass @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" A__ : Tuple = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Optional[Any]: _UpperCamelCase : int = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Any = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def _lowercase ( self , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : int = vqa_pipeline(_snake_case , top_k=1 ) self.assertEqual( _snake_case , [ [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], ] , ) @require_torch def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[int] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question='''How many cats are there?''' , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) _UpperCamelCase : List[Any] = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[Any] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question=_snake_case , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : str = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : Dict = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [[{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''' ) def _lowercase ( self ) -> List[Any]: pass
683
1
'''simple docstring''' class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case ) -> None: _UpperCamelCase : Optional[int] = set_counts _UpperCamelCase : Optional[int] = max(_snake_case ) _UpperCamelCase : Union[str, Any] = len(_snake_case ) _UpperCamelCase : Optional[int] = [1] * num_sets _UpperCamelCase : List[Any] = list(range(_snake_case ) ) def _lowercase ( self , _snake_case , _snake_case ) -> bool: _UpperCamelCase : Any = self.get_parent(_snake_case ) _UpperCamelCase : int = self.get_parent(_snake_case ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] _UpperCamelCase : Union[str, Any] = 0 _UpperCamelCase : str = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 _UpperCamelCase : Optional[int] = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] _UpperCamelCase : Dict = 0 _UpperCamelCase : List[str] = src_parent _UpperCamelCase : Optional[int] = self.set_counts[src_parent] _UpperCamelCase : Any = max(self.max_set , _snake_case ) return True def _lowercase ( self , _snake_case ) -> int: if self.parents[disj_set] == disj_set: return disj_set _UpperCamelCase : List[str] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
683
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers _UpperCAmelCase : Tuple = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) print("""NCCL version:""", torch.cuda.nccl.version()) except ImportError: print("""Torch version:""", None) try: import deepspeed print("""DeepSpeed version:""", deepspeed.__version__) except ImportError: print("""DeepSpeed version:""", None) try: import tensorflow as tf print("""TensorFlow version:""", tf.__version__) print("""TF GPUs available:""", bool(tf.config.list_physical_devices("""GPU"""))) print("""Number of TF GPUs available:""", len(tf.config.list_physical_devices("""GPU"""))) except ImportError: print("""TensorFlow version:""", None)
683
1
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Optional[int] = pytest.mark.integration @pytest.mark.parametrize('''path''' ,['''paws''', '''csv'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: inspect_dataset(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' ,['''accuracy'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: inspect_metric(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[str] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[str]: with pytest.raises(UpperCamelCase ): get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) @pytest.mark.parametrize( '''path, expected''' ,[ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : int = get_dataset_config_names(UpperCamelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' ,[ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Dict = get_dataset_infos(UpperCamelCase ) assert list(infos.keys() ) == expected_configs _UpperCamelCase : Dict = expected_configs[0] assert expected_config in infos _UpperCamelCase : Any = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : List[Any] = get_dataset_infos(UpperCamelCase ) assert expected_config in infos _UpperCamelCase : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: with pytest.raises(UpperCamelCase ): get_dataset_split_names(UpperCamelCase ,config_name=UpperCamelCase )
683
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: return params[f'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase="attention" ) -> List[str]: _UpperCamelCase : Dict = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) _UpperCamelCase : int = k_tmp.reshape(k_tmp.shape[0] ,k_tmp.shape[1] * k_tmp.shape[2] ) _UpperCamelCase : str = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) _UpperCamelCase : Tuple = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] ,o_tmp.shape[2] ) _UpperCamelCase : Any = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) _UpperCamelCase : Optional[int] = q_tmp.reshape(q_tmp.shape[0] ,q_tmp.shape[1] * q_tmp.shape[2] ) _UpperCamelCase : Optional[Any] = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) _UpperCamelCase : List[Any] = v_tmp.reshape(v_tmp.shape[0] ,v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[str]: if split_mlp_wi: _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] _UpperCamelCase : Tuple = params[f'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] _UpperCamelCase : Optional[Any] = (wi_a, wi_a) else: _UpperCamelCase : str = params[f'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: return params[f'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def snake_case__ ( UpperCamelCase ,*, UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ) -> int: _UpperCamelCase : Any = traverse_util.flatten_dict(variables['''target'''] ) _UpperCamelCase : Optional[Any] = {'''/'''.join(UpperCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _UpperCamelCase : str = '''encoder/encoder/mlp/wi_0/kernel''' in old print('''Split MLP:''' ,UpperCamelCase ) _UpperCamelCase : Optional[int] = collections.OrderedDict() # Shared embeddings. _UpperCamelCase : str = old['''token_embedder/embedding'''] # Encoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''attention''' ) _UpperCamelCase : Tuple = layer_norm _UpperCamelCase : int = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : Dict = v.T # Block i, layer 1 (MLP). _UpperCamelCase : Dict = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : int = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,UpperCamelCase ) _UpperCamelCase : Union[str, Any] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Optional[Any] = wi[1].T else: _UpperCamelCase : List[Any] = wi.T _UpperCamelCase : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : Union[str, Any] = tax_relpos_bias_lookup( UpperCamelCase ,UpperCamelCase ,'''encoder''' ).T _UpperCamelCase : List[str] = old['''encoder/encoder_norm/scale'''] if not scalable_attention: _UpperCamelCase : List[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''encoder''' ).T _UpperCamelCase : Optional[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''decoder''' ).T if not is_encoder_only: # Decoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_self_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''self_attention''' ) _UpperCamelCase : int = layer_norm _UpperCamelCase : Union[str, Any] = k.T _UpperCamelCase : Optional[int] = o.T _UpperCamelCase : Dict = q.T _UpperCamelCase : Tuple = v.T # Block i, layer 1 (Cross Attention). _UpperCamelCase : str = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_cross_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Dict = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''encoder_decoder_attention''' ) _UpperCamelCase : Dict = layer_norm _UpperCamelCase : Optional[int] = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : str = v.T # Block i, layer 2 (MLP). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : List[Any] = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,UpperCamelCase ) _UpperCamelCase : List[str] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Union[str, Any] = wi[1].T else: _UpperCamelCase : Dict = wi.T _UpperCamelCase : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : int = tax_relpos_bias_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ).T _UpperCamelCase : Optional[int] = old['''decoder/decoder_norm/scale'''] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _UpperCamelCase : str = old['''decoder/logits_dense/kernel'''].T return new def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Optional[int]: _UpperCamelCase : str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : str = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : int = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''' ) _UpperCamelCase : Any = state_dict['''shared.weight'''] return state_dict def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Any: _UpperCamelCase : List[Any] = checkpoints.load_tax_checkpoint(UpperCamelCase ) _UpperCamelCase : str = convert_tax_to_pytorch( UpperCamelCase ,num_layers=config.num_layers ,is_encoder_only=UpperCamelCase ,scalable_attention=UpperCamelCase ) _UpperCamelCase : Optional[Any] = make_state_dict(UpperCamelCase ,UpperCamelCase ) model.load_state_dict(UpperCamelCase ,strict=UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ,UpperCamelCase = False ,) -> int: _UpperCamelCase : int = MTaConfig.from_json_file(UpperCamelCase ) print(f'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _UpperCamelCase : Optional[int] = UMTaEncoderModel(UpperCamelCase ) else: _UpperCamelCase : Optional[int] = UMTaForConditionalGeneration(UpperCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(UpperCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(UpperCamelCase ) print('''Done''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
683
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : List[Any] = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys _UpperCAmelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
'''simple docstring''' from __future__ import annotations from functools import lru_cache from math import ceil _UpperCAmelCase : int = 100 _UpperCAmelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _UpperCAmelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_00 ) def snake_case__ ( UpperCamelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _UpperCamelCase : set[int] = set() _UpperCamelCase : int _UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def snake_case__ ( UpperCamelCase = 50_00 ) -> int | None: for number_to_partition in range(1 ,UpperCamelCase ): if len(partition(UpperCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
683
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase ( a_ ): """simple docstring""" def __init__( self , *_snake_case , **_snake_case ) -> None: warnings.warn( '''The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PoolFormerImageProcessor instead.''' , _snake_case , ) super().__init__(*_snake_case , **_snake_case )
683
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _UpperCAmelCase : Dict = """bart""" _UpperCAmelCase : List[str] = True @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> int: if LOAD_DENSE_INDEX: _UpperCamelCase : Optional[int] = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _UpperCamelCase : Optional[Any] = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _UpperCamelCase : Tuple = qar_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = (None, None) if MODEL_TYPE == "bart": _UpperCamelCase : Any = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _UpperCamelCase : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _UpperCamelCase : Dict = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _UpperCamelCase : Tuple = sas_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = make_qa_sas_model( model_name='''t5-small''' ,from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' ,device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> List[Any]: if LOAD_DENSE_INDEX: _UpperCamelCase : str = faiss.StandardGpuResources() _UpperCamelCase : Optional[int] = datasets.load_dataset(path='''wiki_snippets''' ,name='''wiki40b_en_100_0''' )['''train'''] _UpperCamelCase : List[str] = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(wikiaab_passages.num_rows, 1_28) ,) _UpperCamelCase : Any = faiss.IndexFlatIP(1_28 ) _UpperCamelCase : str = faiss.index_cpu_to_gpu(UpperCamelCase ,1 ,UpperCamelCase ) wikiaab_gpu_index_flat.add(UpperCamelCase ) # TODO fix for larger GPU else: _UpperCamelCase, _UpperCamelCase : Optional[int] = (None, None) _UpperCamelCase : int = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : List[Any] = datasets.load_dataset('''eli5''' ,name='''LFQA_reddit''' ) _UpperCamelCase : Optional[int] = elia['''train_eli5'''] _UpperCamelCase : Any = np.memmap( '''eli5_questions_reps.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(elia_train.num_rows, 1_28) ) _UpperCamelCase : Optional[Any] = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(UpperCamelCase ) return (elia_train, eli5_train_q_index) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_indexes() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_models() _UpperCAmelCase , _UpperCAmelCase : int = load_train_data() def snake_case__ ( UpperCamelCase ,UpperCamelCase=10 ) -> Optional[Any]: _UpperCamelCase : Optional[int] = embed_questions_for_retrieval([question] ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase, _UpperCamelCase : Optional[Any] = eli5_train_q_index.search(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = [elia_train[int(UpperCamelCase )] for i in I[0]] return nn_examples def snake_case__ ( UpperCamelCase ,UpperCamelCase="wiki40b" ,UpperCamelCase="dense" ,UpperCamelCase=10 ) -> Optional[int]: if source == "none": _UpperCamelCase, _UpperCamelCase : Dict = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _UpperCamelCase, _UpperCamelCase : str = query_qa_dense_index( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else: _UpperCamelCase, _UpperCamelCase : str = query_es_index( UpperCamelCase ,UpperCamelCase ,index_name='''english_wiki40b_snippets_100w''' ,n_results=UpperCamelCase ,) _UpperCamelCase : Optional[int] = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _UpperCamelCase : Optional[Any] = '''question: {} context: {}'''.format(UpperCamelCase ,UpperCamelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda UpperCamelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda UpperCamelCase : None), } ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=64 ,UpperCamelCase=2_56 ,UpperCamelCase=False ,UpperCamelCase=2 ,UpperCamelCase=0.95 ,UpperCamelCase=0.8 ) -> Optional[Any]: with torch.no_grad(): _UpperCamelCase : Any = qa_sas_generate( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,num_answers=1 ,num_beams=UpperCamelCase ,min_len=UpperCamelCase ,max_len=UpperCamelCase ,do_sample=UpperCamelCase ,temp=UpperCamelCase ,top_p=UpperCamelCase ,top_k=UpperCamelCase ,max_input_length=10_24 ,device='''cuda:0''' ,)[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar _UpperCAmelCase : str = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _UpperCAmelCase : Tuple = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _UpperCAmelCase : Dict = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _UpperCAmelCase : List[str] = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _UpperCAmelCase : Optional[int] = st.sidebar.checkbox("""Demo options""") if demo_options: _UpperCAmelCase : List[str] = st.sidebar.selectbox( """""", action_list, index=3, ) _UpperCAmelCase : List[Any] = action_list.index(action_st) _UpperCAmelCase : Tuple = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) _UpperCAmelCase : Optional[Any] = show_type == """Show full text of passages""" else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : str = True _UpperCAmelCase : str = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: _UpperCAmelCase : Optional[Any] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) _UpperCAmelCase : Optional[Any] = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: _UpperCAmelCase : Dict = """wiki40b""" _UpperCAmelCase : str = """dense""" _UpperCAmelCase : List[str] = """beam""" _UpperCAmelCase : Dict = 2 _UpperCAmelCase : List[str] = 64 _UpperCAmelCase : List[Any] = 256 _UpperCAmelCase : Tuple = None _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : int = st.sidebar.checkbox("""Generation options""") if generate_options: _UpperCAmelCase : Union[str, Any] = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) _UpperCAmelCase : Dict = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) _UpperCAmelCase : List[Any] = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": _UpperCAmelCase : List[str] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[int] = None # start main text _UpperCAmelCase : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _UpperCAmelCase : int = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": _UpperCAmelCase : Any = st.text_input("""Enter your question here:""", """""") else: _UpperCAmelCase : Tuple = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": _UpperCAmelCase , _UpperCAmelCase : str = make_support(question, source=wiki_source, method="""dense""", n_results=10) _UpperCAmelCase , _UpperCAmelCase : List[Any] = make_support(question, source=wiki_source, method="""sparse""", n_results=10) _UpperCAmelCase : int = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _UpperCAmelCase : int = support_list[:10] _UpperCAmelCase : Tuple = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _UpperCAmelCase , _UpperCAmelCase : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): _UpperCAmelCase : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) _UpperCAmelCase : List[Any] = res[1].strip() if sec_titles == "": _UpperCAmelCase : Optional[int] = """[{}]({})""".format(res[0], wiki_url) else: _UpperCAmelCase : Optional[int] = sec_titles.split(""" & """) _UpperCAmelCase : Tuple = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: _UpperCAmelCase : Dict = find_nearest_training(question) _UpperCAmelCase : List[Any] = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) _UpperCAmelCase : List[Any] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) _UpperCAmelCase : List[Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
683
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Any = { """kssteven/ibert-roberta-base""": """https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json""", """kssteven/ibert-roberta-large""": """https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json""", """kssteven/ibert-roberta-large-mnli""": ( """https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json""" ), } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Tuple = 'ibert' def __init__( self , _snake_case=30522 , _snake_case=768 , _snake_case=12 , _snake_case=12 , _snake_case=3072 , _snake_case="gelu" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=2 , _snake_case=0.02 , _snake_case=1E-12 , _snake_case=1 , _snake_case=0 , _snake_case=2 , _snake_case="absolute" , _snake_case=False , _snake_case="none" , **_snake_case , ) -> str: super().__init__(pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , **_snake_case ) _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Any = num_hidden_layers _UpperCamelCase : Optional[Any] = num_attention_heads _UpperCamelCase : str = hidden_act _UpperCamelCase : List[str] = intermediate_size _UpperCamelCase : Dict = hidden_dropout_prob _UpperCamelCase : Optional[int] = attention_probs_dropout_prob _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : Any = type_vocab_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : str = layer_norm_eps _UpperCamelCase : Dict = position_embedding_type _UpperCamelCase : Optional[int] = quant_mode _UpperCamelCase : Union[str, Any] = force_dequant class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCamelCase : Union[str, Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCamelCase : Optional[int] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
683
'''simple docstring''' from collections.abc import Iterable from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> Optional[int]: _UpperCamelCase : int = value _UpperCamelCase : Node | None = None # Added in order to delete a node easier _UpperCamelCase : Node | None = None _UpperCamelCase : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> List[Any]: _UpperCamelCase : str = root def __str__( self ) -> str: return str(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if new_children is not None: # reset its kids _UpperCamelCase : Union[str, Any] = node.parent if node.parent is not None: # reset its parent if self.is_right(_snake_case ): # If it is the right children _UpperCamelCase : str = new_children else: _UpperCamelCase : Any = new_children else: _UpperCamelCase : Any = new_children def _lowercase ( self , _snake_case ) -> bool: if node.parent and node.parent.right: return node == node.parent.right return False def _lowercase ( self ) -> bool: return self.root is None def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : List[Any] = Node(_snake_case ) # create a new Node if self.empty(): # if Tree is empty _UpperCamelCase : Optional[Any] = new_node # set its root else: # Tree is not empty _UpperCamelCase : int = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: _UpperCamelCase : Union[str, Any] = new_node # We insert the new node in a leaf break else: _UpperCamelCase : Union[str, Any] = parent_node.left else: if parent_node.right is None: _UpperCamelCase : Any = new_node break else: _UpperCamelCase : str = parent_node.right _UpperCamelCase : Any = parent_node def _lowercase ( self , *_snake_case ) -> None: for value in values: self.__insert(_snake_case ) def _lowercase ( self , _snake_case ) -> Node | None: if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: _UpperCamelCase : List[str] = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: _UpperCamelCase : Optional[Any] = node.left if value < node.value else node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: if self.root is None: return None _UpperCamelCase : Dict = self.root if not self.empty(): while node.right is not None: _UpperCamelCase : Tuple = node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: _UpperCamelCase : Optional[Any] = self.root if self.root is None: return None if not self.empty(): _UpperCamelCase : Optional[int] = self.root while node.left is not None: _UpperCamelCase : List[str] = node.left return node def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : str = self.search(_snake_case ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_snake_case , _snake_case ) elif node.left is None: # Has only right children self.__reassign_nodes(_snake_case , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_snake_case , node.left ) else: _UpperCamelCase : List[str] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore _UpperCamelCase : int = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _lowercase ( self , _snake_case ) -> Iterable: if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _lowercase ( self , _snake_case=None ) -> Any: if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if node: self.inorder(_snake_case , node.left ) arr.append(node.value ) self.inorder(_snake_case , node.right ) def _lowercase ( self , _snake_case , _snake_case ) -> int: _UpperCamelCase : list[int] = [] self.inorder(_snake_case , _snake_case ) # append all values to list using inorder traversal return arr[k - 1] def snake_case__ ( UpperCamelCase ) -> list[Node]: _UpperCamelCase : int = [] if curr_node is not None: _UpperCamelCase : Any = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def snake_case__ ( ) -> None: _UpperCamelCase : Any = (8, 3, 6, 1, 10, 14, 13, 4, 7) _UpperCamelCase : Tuple = BinarySearchTree() for i in testlist: t.insert(UpperCamelCase ) # Prints all the elements of the list in order traversal print(UpperCamelCase ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' ,t.get_max().value ) # type: ignore print('''Min Value: ''' ,t.get_min().value ) # type: ignore for i in testlist: t.remove(UpperCamelCase ) print(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
683
1
'''simple docstring''' from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Tuple = 'EncodecFeatureExtractor' A__ : str = ('T5Tokenizer', 'T5TokenizerFast') def __init__( self , _snake_case , _snake_case ) -> Tuple: super().__init__(_snake_case , _snake_case ) _UpperCamelCase : Tuple = self.feature_extractor _UpperCamelCase : List[str] = False def _lowercase ( self , _snake_case=None , _snake_case=None , _snake_case=True ) -> Any: return self.tokenizer.get_decoder_prompt_ids(task=_snake_case , language=_snake_case , no_timestamps=_snake_case ) def __call__( self , *_snake_case , **_snake_case ) -> str: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_snake_case , **_snake_case ) _UpperCamelCase : Any = kwargs.pop('''audio''' , _snake_case ) _UpperCamelCase : int = kwargs.pop('''sampling_rate''' , _snake_case ) _UpperCamelCase : List[str] = kwargs.pop('''text''' , _snake_case ) if len(_snake_case ) > 0: _UpperCamelCase : Tuple = args[0] _UpperCamelCase : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if text is not None: _UpperCamelCase : Union[str, Any] = self.tokenizer(_snake_case , **_snake_case ) if audio is not None: _UpperCamelCase : str = self.feature_extractor(_snake_case , *_snake_case , sampling_rate=_snake_case , **_snake_case ) if audio is None: return inputs elif text is None: return audio_inputs else: _UpperCamelCase : str = audio_inputs['''input_values'''] if "padding_mask" in audio_inputs: _UpperCamelCase : Optional[int] = audio_inputs['''padding_mask'''] return inputs def _lowercase ( self , *_snake_case , **_snake_case ) -> int: _UpperCamelCase : Any = kwargs.pop('''audio''' , _snake_case ) _UpperCamelCase : int = kwargs.pop('''padding_mask''' , _snake_case ) if len(_snake_case ) > 0: _UpperCamelCase : Any = args[0] _UpperCamelCase : str = args[1:] if audio_values is not None: return self._decode_audio(_snake_case , padding_mask=_snake_case ) else: return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> int: return self.tokenizer.decode(*_snake_case , **_snake_case ) def _lowercase ( self , _snake_case , _snake_case = None ) -> List[np.ndarray]: _UpperCamelCase : Dict = to_numpy(_snake_case ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = audio_values.shape if padding_mask is None: return list(_snake_case ) _UpperCamelCase : List[str] = to_numpy(_snake_case ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _UpperCamelCase : Optional[int] = seq_len - padding_mask.shape[-1] _UpperCamelCase : List[Any] = 1 - self.feature_extractor.padding_value _UpperCamelCase : Dict = np.pad(_snake_case , ((0, 0), (0, difference)) , '''constant''' , constant_values=_snake_case ) _UpperCamelCase : List[str] = audio_values.tolist() for i in range(_snake_case ): _UpperCamelCase : Any = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _UpperCamelCase : List[Any] = sliced_audio.reshape(_snake_case , -1 ) return audio_values
683
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off _UpperCAmelCase : Dict = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 10563, 10786, 11420, 11709, 11907, 13163, 13697, 13700, 14808, 15306, 16410, 16791, 17992, 19203, 19510, 20724, 22305, 22935, 27007, 30109, 30420, 33409, 34949, 40283, 40493, 40549, 47282, 49146, 50257, 50359, 50360, 50361 ] _UpperCAmelCase : int = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362 ] class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = 'whisper' A__ : Tuple = ['past_key_values'] A__ : Optional[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , _snake_case=51865 , _snake_case=80 , _snake_case=6 , _snake_case=4 , _snake_case=6 , _snake_case=4 , _snake_case=1536 , _snake_case=1536 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=50257 , _snake_case=True , _snake_case=True , _snake_case="gelu" , _snake_case=256 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.02 , _snake_case=False , _snake_case=1500 , _snake_case=448 , _snake_case=50256 , _snake_case=50256 , _snake_case=50256 , _snake_case=None , _snake_case=[220, 50256] , _snake_case=False , _snake_case=256 , _snake_case=False , _snake_case=0.05 , _snake_case=10 , _snake_case=2 , _snake_case=0.0 , _snake_case=10 , _snake_case=0 , _snake_case=7 , **_snake_case , ) -> Any: _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Union[str, Any] = num_mel_bins _UpperCamelCase : List[str] = d_model _UpperCamelCase : str = encoder_layers _UpperCamelCase : Optional[int] = encoder_attention_heads _UpperCamelCase : str = decoder_layers _UpperCamelCase : Tuple = decoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : Optional[int] = encoder_ffn_dim _UpperCamelCase : Any = dropout _UpperCamelCase : Optional[Any] = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : int = activation_function _UpperCamelCase : List[Any] = init_std _UpperCamelCase : Optional[int] = encoder_layerdrop _UpperCamelCase : str = decoder_layerdrop _UpperCamelCase : List[str] = use_cache _UpperCamelCase : Optional[Any] = encoder_layers _UpperCamelCase : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : List[str] = max_source_positions _UpperCamelCase : Optional[Any] = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. _UpperCamelCase : str = classifier_proj_size _UpperCamelCase : List[str] = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCamelCase : int = apply_spec_augment _UpperCamelCase : str = mask_time_prob _UpperCamelCase : int = mask_time_length _UpperCamelCase : List[Any] = mask_time_min_masks _UpperCamelCase : List[str] = mask_feature_prob _UpperCamelCase : Optional[int] = mask_feature_length _UpperCamelCase : Union[str, Any] = mask_feature_min_masks _UpperCamelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , is_encoder_decoder=_snake_case , decoder_start_token_id=_snake_case , suppress_tokens=_snake_case , begin_suppress_tokens=_snake_case , **_snake_case , ) class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: _UpperCamelCase : Dict = OrderedDict( [ ('''input_features''', {0: '''batch''', 1: '''feature_size''', 2: '''encoder_sequence'''}), ] ) if self.use_past: _UpperCamelCase : Tuple = {0: '''batch'''} else: _UpperCamelCase : Dict = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_snake_case , direction='''inputs''' ) return common_inputs def _lowercase ( self , _snake_case , _snake_case = -1 , _snake_case = -1 , _snake_case = False , _snake_case = None , _snake_case = 22050 , _snake_case = 5.0 , _snake_case = 220 , ) -> Mapping[str, Any]: _UpperCamelCase : Optional[int] = OrderedDict() _UpperCamelCase : Tuple = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_snake_case , framework=_snake_case , sampling_rate=_snake_case , time_duration=_snake_case , frequency=_snake_case , ) _UpperCamelCase : int = encoder_inputs['''input_features'''].shape[2] _UpperCamelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length _UpperCamelCase : str = super().generate_dummy_inputs( preprocessor.tokenizer , _snake_case , _snake_case , _snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = encoder_inputs.pop('''input_features''' ) _UpperCamelCase : Dict = decoder_inputs.pop('''decoder_input_ids''' ) if "past_key_values" in decoder_inputs: _UpperCamelCase : List[str] = decoder_inputs.pop('''past_key_values''' ) return dummy_inputs @property def _lowercase ( self ) -> float: return 1E-3
683
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_longt5""": ["""LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongT5Config""", """LongT5OnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongT5EncoderModel""", """LongT5ForConditionalGeneration""", """LongT5Model""", """LongT5PreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """FlaxLongT5ForConditionalGeneration""", """FlaxLongT5Model""", """FlaxLongT5PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
'''simple docstring''' import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser( description=( """Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""roberta""", choices=["""roberta""", """gpt2"""]) parser.add_argument("""--model_name""", default="""roberta-large""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_roberta_048131723.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") _UpperCAmelCase : int = parser.parse_args() if args.model_type == "roberta": _UpperCAmelCase : Union[str, Any] = RobertaForMaskedLM.from_pretrained(args.model_name) _UpperCAmelCase : int = """roberta""" elif args.model_type == "gpt2": _UpperCAmelCase : Optional[int] = GPTaLMHeadModel.from_pretrained(args.model_name) _UpperCAmelCase : Optional[int] = """transformer""" _UpperCAmelCase : Tuple = model.state_dict() _UpperCAmelCase : int = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: _UpperCAmelCase : Optional[Any] = state_dict[f"""{prefix}.{param_name}"""] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: _UpperCAmelCase : Tuple = f"""{prefix}.embeddings.{w}.weight""" _UpperCAmelCase : Optional[Any] = state_dict[param_name] for w in ["weight", "bias"]: _UpperCAmelCase : Union[str, Any] = f"""{prefix}.embeddings.LayerNorm.{w}""" _UpperCAmelCase : str = state_dict[param_name] # Transformer Blocks # _UpperCAmelCase : Dict = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: _UpperCAmelCase : str = state_dict[ f"""{prefix}.h.{teacher_idx}.{layer}.{w}""" ] _UpperCAmelCase : Any = state_dict[f"""{prefix}.h.{teacher_idx}.attn.bias"""] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: _UpperCAmelCase : Optional[Any] = state_dict[ f"""{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}""" ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: _UpperCAmelCase : Dict = state_dict[f"""{layer}"""] if args.vocab_transform: for w in ["weight", "bias"]: _UpperCAmelCase : int = state_dict[f"""lm_head.dense.{w}"""] _UpperCAmelCase : int = state_dict[f"""lm_head.layer_norm.{w}"""] elif args.model_type == "gpt2": for w in ["weight", "bias"]: _UpperCAmelCase : List[str] = state_dict[f"""{prefix}.ln_f.{w}"""] _UpperCAmelCase : Any = state_dict["""lm_head.weight"""] print(f"""N layers selected for distillation: {std_idx}""") print(f"""Number of params transferred for distillation: {len(compressed_sd.keys())}""") print(f"""Save transferred checkpoint to {args.dump_checkpoint}.""") torch.save(compressed_sd, args.dump_checkpoint)
683
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _UpperCAmelCase : Dict = logging.get_logger(__name__) class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = ['pixel_values'] def __init__( self , _snake_case = True , _snake_case = 32 , _snake_case=PILImageResampling.BILINEAR , _snake_case = True , **_snake_case , ) -> None: _UpperCamelCase : Union[str, Any] = do_resize _UpperCamelCase : Optional[Any] = do_rescale _UpperCamelCase : str = size_divisor _UpperCamelCase : Optional[int] = resample super().__init__(**_snake_case ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case = None , **_snake_case ) -> np.ndarray: _UpperCamelCase, _UpperCamelCase : Any = get_image_size(_snake_case ) # Rounds the height and width down to the closest multiple of size_divisor _UpperCamelCase : str = height // size_divisor * size_divisor _UpperCamelCase : List[Any] = width // size_divisor * size_divisor _UpperCamelCase : Tuple = resize(_snake_case , (new_h, new_w) , resample=_snake_case , data_format=_snake_case , **_snake_case ) return image def _lowercase ( self , _snake_case , _snake_case , _snake_case = None , **_snake_case ) -> np.ndarray: return rescale(image=_snake_case , scale=_snake_case , data_format=_snake_case , **_snake_case ) def _lowercase ( self , _snake_case , _snake_case = None , _snake_case = None , _snake_case=None , _snake_case = None , _snake_case = None , _snake_case = ChannelDimension.FIRST , **_snake_case , ) -> BatchFeature: _UpperCamelCase : Optional[int] = do_resize if do_resize is not None else self.do_resize _UpperCamelCase : List[str] = do_rescale if do_rescale is not None else self.do_rescale _UpperCamelCase : Dict = size_divisor if size_divisor is not None else self.size_divisor _UpperCamelCase : Tuple = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''' ) _UpperCamelCase : Optional[Any] = make_list_of_images(_snake_case ) if not valid_images(_snake_case ): raise ValueError('''Invalid image(s)''' ) # All transformations expect numpy arrays. _UpperCamelCase : Optional[int] = [to_numpy_array(_snake_case ) for img in images] if do_resize: _UpperCamelCase : List[str] = [self.resize(_snake_case , size_divisor=_snake_case , resample=_snake_case ) for image in images] if do_rescale: _UpperCamelCase : str = [self.rescale(_snake_case , scale=1 / 255 ) for image in images] _UpperCamelCase : List[str] = [to_channel_dimension_format(_snake_case , _snake_case ) for image in images] _UpperCamelCase : Tuple = {'''pixel_values''': images} return BatchFeature(data=_snake_case , tensor_type=_snake_case )
683
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = None _UpperCamelCase : int = 20 _UpperCamelCase : Any = self._get_uniform_logits(batch_size=2 , length=_snake_case ) # tweak scores to not be uniform anymore _UpperCamelCase : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch _UpperCamelCase : Dict = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax _UpperCamelCase : Any = jax.nn.softmax(_snake_case , axis=-1 ) _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : List[str] = FlaxTemperatureLogitsWarper(temperature=1.3 ) _UpperCamelCase : List[str] = jax.nn.softmax(temp_dist_warper_sharper(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) _UpperCamelCase : str = jax.nn.softmax(temp_dist_warper_smoother(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _lowercase ( self ) -> Any: _UpperCamelCase : List[Any] = None _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : Any = 2 # create ramp distribution _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() _UpperCamelCase : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case _UpperCamelCase : Optional[int] = 5 _UpperCamelCase : str = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) _UpperCamelCase : Union[str, Any] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, length) ).copy() _UpperCamelCase : Optional[Any] = top_k_warp_safety_check(_snake_case , _snake_case , cur_len=_snake_case ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = None _UpperCamelCase : Any = 10 _UpperCamelCase : List[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) _UpperCamelCase : Tuple = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) _UpperCamelCase : List[str] = FlaxTopPLogitsWarper(0.8 ) _UpperCamelCase : Dict = np.exp(top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 _UpperCamelCase : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # check edge cases with negative and extreme logits _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme _UpperCamelCase : Tuple = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept _UpperCamelCase : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) _UpperCamelCase : Dict = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = 20 _UpperCamelCase : Optional[int] = 4 _UpperCamelCase : int = 0 _UpperCamelCase : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) # check that min length is applied at length 5 _UpperCamelCase : Any = ids_tensor((batch_size, 20) , vocab_size=20 ) _UpperCamelCase : int = 5 _UpperCamelCase : List[Any] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 _UpperCamelCase : Optional[int] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = 15 _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = 20 _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) # check that all scores are -inf except the bos_token_id score _UpperCamelCase : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) _UpperCamelCase : Optional[int] = 1 _UpperCamelCase : str = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : str = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 _UpperCamelCase : List[str] = 3 _UpperCamelCase : Tuple = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 20 _UpperCamelCase : Tuple = 4 _UpperCamelCase : Any = 0 _UpperCamelCase : str = 5 _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) # check that all scores are -inf except the eos_token_id when max_length is reached _UpperCamelCase : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) _UpperCamelCase : Dict = 4 _UpperCamelCase : Dict = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : int = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Any = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 4 _UpperCamelCase : Optional[Any] = 10 _UpperCamelCase : Dict = 15 _UpperCamelCase : Union[str, Any] = 2 _UpperCamelCase : Optional[Any] = 1 _UpperCamelCase : List[Any] = 15 # dummy input_ids and scores _UpperCamelCase : Optional[int] = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Any = input_ids.copy() _UpperCamelCase : int = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Tuple = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Optional[int] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : int = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : List[str] = 10 # no processor list _UpperCamelCase : Dict = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) # with processor list _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : Optional[Any] = processor(_snake_case , _snake_case , cur_len=_snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = 4 _UpperCamelCase : int = 10 _UpperCamelCase : List[Any] = 15 _UpperCamelCase : Dict = 2 _UpperCamelCase : Tuple = 1 _UpperCamelCase : Optional[int] = 15 # dummy input_ids and scores _UpperCamelCase : Tuple = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Optional[Any] = input_ids.copy() _UpperCamelCase : List[str] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[int] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : int = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : Union[str, Any] = 10 # no processor list def run_no_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : List[Any] = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) return scores # with processor list def run_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : List[str] = processor(_snake_case , _snake_case , cur_len=_snake_case ) return scores _UpperCamelCase : Dict = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jitted_run_no_processor_list(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = jitted_run_processor_list(_snake_case , _snake_case , _snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
683
1
'''simple docstring''' import numpy as np from PIL import Image def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> np.ndarray: _UpperCamelCase : Any = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) _UpperCamelCase : Optional[Any] = 0 _UpperCamelCase : Union[str, Any] = 0 _UpperCamelCase : Dict = 0 _UpperCamelCase : Any = 0 # compute the shape of the output matrix _UpperCamelCase : List[str] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape _UpperCamelCase : Optional[int] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix _UpperCamelCase : Tuple = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _UpperCamelCase : List[str] = 0 _UpperCamelCase : int = 0 return updated_arr def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> np.ndarray: _UpperCamelCase : Optional[Any] = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) _UpperCamelCase : List[str] = 0 _UpperCamelCase : Union[str, Any] = 0 _UpperCamelCase : int = 0 _UpperCamelCase : int = 0 # compute the shape of the output matrix _UpperCamelCase : int = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape _UpperCamelCase : Union[str, Any] = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix _UpperCamelCase : Optional[int] = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _UpperCamelCase : Optional[Any] = 0 _UpperCamelCase : Optional[int] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="""avgpooling""", verbose=True) # Loading the image _UpperCAmelCase : Optional[Any] = Image.open("""path_to_image""") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
683
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Optional[int] = pytest.mark.integration @pytest.mark.parametrize('''path''' ,['''paws''', '''csv'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: inspect_dataset(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' ,['''accuracy'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: inspect_metric(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[str] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[str]: with pytest.raises(UpperCamelCase ): get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) @pytest.mark.parametrize( '''path, expected''' ,[ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : int = get_dataset_config_names(UpperCamelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' ,[ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Dict = get_dataset_infos(UpperCamelCase ) assert list(infos.keys() ) == expected_configs _UpperCamelCase : Dict = expected_configs[0] assert expected_config in infos _UpperCamelCase : Any = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : List[Any] = get_dataset_infos(UpperCamelCase ) assert expected_config in infos _UpperCamelCase : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: with pytest.raises(UpperCamelCase ): get_dataset_split_names(UpperCamelCase ,config_name=UpperCamelCase )
683
1
'''simple docstring''' from math import sqrt def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : List[str] = 0 for i in range(1 ,int(sqrt(UpperCamelCase ) + 1 ) ): if n % i == 0 and i != sqrt(UpperCamelCase ): total += i + n // i elif i == sqrt(UpperCamelCase ): total += i return total - n def snake_case__ ( UpperCamelCase = 1_00_00 ) -> int: _UpperCamelCase : Optional[int] = sum( i for i in range(1 ,UpperCamelCase ) if sum_of_divisors(sum_of_divisors(UpperCamelCase ) ) == i and sum_of_divisors(UpperCamelCase ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
683
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ) -> Dict: torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return model @property def _lowercase ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , cross_attention_dim=10 , ) return model @property def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') , up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') , ) _UpperCamelCase : int = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return vqvae, unet @slow def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Tuple = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _UpperCamelCase : int = DDPMScheduler() _UpperCamelCase : Optional[int] = AudioDiffusionPipeline(vqvae=_snake_case , unet=self.dummy_unet , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case , steps=4 ) _UpperCamelCase : Union[str, Any] = output.audios[0] _UpperCamelCase : Union[str, Any] = output.images[0] _UpperCamelCase : str = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : int = pipe(generator=_snake_case , steps=4 , return_dict=_snake_case ) _UpperCamelCase : int = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _UpperCamelCase : List[str] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : List[str] = np.frombuffer(image_from_tuple.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : int = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : str = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _UpperCamelCase : Dict = DDIMScheduler() _UpperCamelCase : str = self.dummy_vqvae_and_unet _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : Optional[Any] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _UpperCamelCase : Optional[Any] = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Tuple = pipe(raw_audio=_snake_case , generator=_snake_case , start_step=5 , steps=10 ) _UpperCamelCase : List[str] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _UpperCamelCase : Any = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Tuple = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : Any = self.dummy_unet_condition _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_snake_case , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : Union[str, Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : int = torch.rand((1, 1, 10) ) _UpperCamelCase : Optional[Any] = pipe(generator=_snake_case , encoding=_snake_case ) _UpperCamelCase : Dict = output.images[0] _UpperCamelCase : Dict = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Any = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = torch_device _UpperCamelCase : int = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) _UpperCamelCase : str = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case ) _UpperCamelCase : List[Any] = output.audios[0] _UpperCamelCase : List[Any] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _UpperCamelCase : Union[str, Any] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Union[str, Any] = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> bool: _UpperCamelCase : Optional[Any] = (1 + 24 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def snake_case__ ( UpperCamelCase = 50_00 ) -> int: _UpperCamelCase : Dict = [(i * (3 * i - 1)) // 2 for i in range(1 ,UpperCamelCase )] for i, pentagonal_i in enumerate(UpperCamelCase ): for j in range(UpperCamelCase ,len(UpperCamelCase ) ): _UpperCamelCase : Tuple = pentagonal_nums[j] _UpperCamelCase : int = pentagonal_i + pentagonal_j _UpperCamelCase : List[Any] = pentagonal_j - pentagonal_i if is_pentagonal(UpperCamelCase ) and is_pentagonal(UpperCamelCase ): return b return -1 if __name__ == "__main__": print(f"""{solution() = }""")
683
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_longt5""": ["""LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongT5Config""", """LongT5OnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongT5EncoderModel""", """LongT5ForConditionalGeneration""", """LongT5Model""", """LongT5PreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """FlaxLongT5ForConditionalGeneration""", """FlaxLongT5Model""", """FlaxLongT5PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
1
'''simple docstring''' import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : List[str] = LxmertTokenizer A__ : Union[str, Any] = LxmertTokenizerFast A__ : Any = True A__ : Union[str, Any] = True def _lowercase ( self ) -> Dict: super().setUp() _UpperCamelCase : Optional[int] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] _UpperCamelCase : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def _lowercase ( self , _snake_case ) -> Optional[int]: _UpperCamelCase : List[str] = '''UNwant\u00E9d,running''' _UpperCamelCase : Dict = '''unwanted, running''' return input_text, output_text def _lowercase ( self ) -> int: _UpperCamelCase : List[str] = self.tokenizer_class(self.vocab_file ) _UpperCamelCase : Any = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(_snake_case , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_snake_case ) , [7, 4, 5, 10, 8, 9] ) def _lowercase ( self ) -> List[str]: if not self.test_rust_tokenizer: return _UpperCamelCase : Optional[Any] = self.get_tokenizer() _UpperCamelCase : List[Any] = self.get_rust_tokenizer() _UpperCamelCase : Union[str, Any] = '''I was born in 92000, and this is falsé.''' _UpperCamelCase : List[str] = tokenizer.tokenize(_snake_case ) _UpperCamelCase : Optional[Any] = rust_tokenizer.tokenize(_snake_case ) self.assertListEqual(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = tokenizer.encode(_snake_case , add_special_tokens=_snake_case ) _UpperCamelCase : Any = rust_tokenizer.encode(_snake_case , add_special_tokens=_snake_case ) self.assertListEqual(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = self.get_rust_tokenizer() _UpperCamelCase : List[Any] = tokenizer.encode(_snake_case ) _UpperCamelCase : Tuple = rust_tokenizer.encode(_snake_case ) self.assertListEqual(_snake_case , _snake_case )
683
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } _UpperCAmelCase : Any = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : Tuple = DistilBertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> int: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : int = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = do_lower_case _UpperCamelCase : Dict = strip_accents _UpperCamelCase : List[Any] = tokenize_chinese_chars _UpperCamelCase : Tuple = normalizer_class(**_snake_case ) _UpperCamelCase : Dict = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[int]: _UpperCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Optional[Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
1
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Optional[int] = TextToVideoSDPipeline A__ : Dict = TEXT_TO_IMAGE_PARAMS A__ : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. A__ : str = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) _UpperCamelCase : Tuple = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_snake_case , set_alpha_to_one=_snake_case , ) torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _UpperCamelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) _UpperCamelCase : Any = CLIPTextModel(_snake_case ) _UpperCamelCase : Any = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _UpperCamelCase : Dict = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def _lowercase ( self , _snake_case , _snake_case=0 ) -> Any: if str(_snake_case ).startswith('''mps''' ): _UpperCamelCase : str = torch.manual_seed(_snake_case ) else: _UpperCamelCase : str = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) _UpperCamelCase : List[str] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def _lowercase ( self ) -> int: _UpperCamelCase : Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Any = self.get_dummy_components() _UpperCamelCase : List[str] = TextToVideoSDPipeline(**_snake_case ) _UpperCamelCase : List[str] = sd_pipe.to(_snake_case ) sd_pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Optional[Any] = self.get_dummy_inputs(_snake_case ) _UpperCamelCase : List[str] = '''np''' _UpperCamelCase : Tuple = sd_pipe(**_snake_case ).frames _UpperCamelCase : Optional[int] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _UpperCamelCase : List[Any] = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self ) -> str: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=_snake_case , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def _lowercase ( self ) -> Any: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=_snake_case , expected_max_diff=1E-2 ) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def _lowercase ( self ) -> Optional[int]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def _lowercase ( self ) -> Union[str, Any]: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''' ) def _lowercase ( self ) -> Optional[int]: pass def _lowercase ( self ) -> Union[str, Any]: return super().test_progress_bar() @slow @skip_mps class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''' ) _UpperCamelCase : Optional[int] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) _UpperCamelCase : Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _UpperCamelCase : Optional[Any] = pipe.to('''cuda''' ) _UpperCamelCase : List[str] = '''Spiderman is surfing''' _UpperCamelCase : str = torch.Generator(device='''cpu''' ).manual_seed(0 ) _UpperCamelCase : Tuple = pipe(_snake_case , generator=_snake_case , num_inference_steps=25 , output_type='''pt''' ).frames _UpperCamelCase : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def _lowercase ( self ) -> str: _UpperCamelCase : str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''' ) _UpperCamelCase : Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) _UpperCamelCase : int = pipe.to('''cuda''' ) _UpperCamelCase : Tuple = '''Spiderman is surfing''' _UpperCamelCase : Optional[Any] = torch.Generator(device='''cpu''' ).manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipe(_snake_case , generator=_snake_case , num_inference_steps=2 , output_type='''pt''' ).frames _UpperCamelCase : List[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
683
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> list: _UpperCamelCase : Any = False while is_sorted is False: # Until all the indices are traversed keep looping _UpperCamelCase : List[str] = True for i in range(0 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Dict = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : int = False for i in range(1 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Optional[Any] = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : Optional[int] = False return input_list if __name__ == "__main__": print("""Enter list to be sorted""") _UpperCAmelCase : Optional[int] = [int(x) for x in input().split()] # inputing elements of the list in one line _UpperCAmelCase : Union[str, Any] = odd_even_sort(input_list) print("""The sorted list is""") print(sorted_list)
683
1
'''simple docstring''' from collections import defaultdict def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> bool: _UpperCamelCase : str = first_str.lower().strip() _UpperCamelCase : Dict = second_str.lower().strip() # Remove whitespace _UpperCamelCase : Any = first_str.replace(''' ''' ,'''''' ) _UpperCamelCase : Dict = second_str.replace(''' ''' ,'''''' ) # Strings of different lengths are not anagrams if len(UpperCamelCase ) != len(UpperCamelCase ): return False # Default values for count should be 0 _UpperCamelCase : defaultdict[str, int] = defaultdict(UpperCamelCase ) # For each character in input strings, # increment count in the corresponding for i in range(len(UpperCamelCase ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() _UpperCAmelCase : int = input("""Enter the first string """).strip() _UpperCAmelCase : Optional[int] = input("""Enter the second string """).strip() _UpperCAmelCase : List[Any] = check_anagrams(input_a, input_b) print(f"""{input_a} and {input_b} are {'' if status else 'not '}anagrams.""")
683
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = checkpoint _UpperCamelCase : int = {} _UpperCamelCase : int = vae_state_dict['''encoder.conv_in.weight'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_in.bias'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_out.weight'''] _UpperCamelCase : Any = vae_state_dict['''encoder.conv_out.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''encoder.norm_out.weight'''] _UpperCamelCase : str = vae_state_dict['''encoder.norm_out.bias'''] _UpperCamelCase : str = vae_state_dict['''decoder.conv_in.weight'''] _UpperCamelCase : List[Any] = vae_state_dict['''decoder.conv_in.bias'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.weight'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.bias'''] _UpperCamelCase : int = vae_state_dict['''decoder.norm_out.weight'''] _UpperCamelCase : Dict = vae_state_dict['''decoder.norm_out.bias'''] _UpperCamelCase : Optional[int] = vae_state_dict['''quant_conv.weight'''] _UpperCamelCase : int = vae_state_dict['''quant_conv.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''post_quant_conv.weight'''] _UpperCamelCase : Optional[int] = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only _UpperCamelCase : Optional[int] = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) _UpperCamelCase : Tuple = { layer_id: [key for key in vae_state_dict if f'''down.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } # Retrieves the keys for the decoder up blocks only _UpperCamelCase : Any = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) _UpperCamelCase : int = { layer_id: [key for key in vae_state_dict if f'''up.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } for i in range(UpperCamelCase ): _UpperCamelCase : Any = [key for key in down_blocks[i] if f'''down.{i}''' in key and f'''down.{i}.downsample''' not in key] if f'''encoder.down.{i}.downsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Optional[int] = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.weight''' ) _UpperCamelCase : Dict = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.bias''' ) _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Union[str, Any] = {'''old''': f'''down.{i}.block''', '''new''': f'''down_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[str] = [key for key in vae_state_dict if '''encoder.mid.block''' in key] _UpperCamelCase : Tuple = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : Optional[int] = [key for key in mid_resnets if f'''encoder.mid.block_{i}''' in key] _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Tuple = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] _UpperCamelCase : List[str] = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) for i in range(UpperCamelCase ): _UpperCamelCase : Union[str, Any] = num_up_blocks - 1 - i _UpperCamelCase : Optional[int] = [ key for key in up_blocks[block_id] if f'''up.{block_id}''' in key and f'''up.{block_id}.upsample''' not in key ] if f'''decoder.up.{block_id}.upsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Tuple = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.weight''' ] _UpperCamelCase : Any = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.bias''' ] _UpperCamelCase : Any = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': f'''up.{block_id}.block''', '''new''': f'''up_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[Any] = [key for key in vae_state_dict if '''decoder.mid.block''' in key] _UpperCamelCase : Optional[Any] = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : int = [key for key in mid_resnets if f'''decoder.mid.block_{i}''' in key] _UpperCamelCase : Optional[int] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Optional[Any] = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] _UpperCamelCase : Tuple = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Dict = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) return new_checkpoint def snake_case__ ( UpperCamelCase ,UpperCamelCase ,) -> List[str]: # Only support V1 _UpperCamelCase : Tuple = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) _UpperCamelCase : List[Any] = io.BytesIO(r.content ) _UpperCamelCase : Optional[int] = OmegaConf.load(UpperCamelCase ) _UpperCamelCase : str = 5_12 _UpperCamelCase : int = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open _UpperCamelCase : str = {} with safe_open(UpperCamelCase ,framework='''pt''' ,device='''cpu''' ) as f: for key in f.keys(): _UpperCamelCase : Union[str, Any] = f.get_tensor(UpperCamelCase ) else: _UpperCamelCase : str = torch.load(UpperCamelCase ,map_location=UpperCamelCase )['''state_dict'''] # Convert the VAE model. _UpperCamelCase : Dict = create_vae_diffusers_config(UpperCamelCase ,image_size=UpperCamelCase ) _UpperCamelCase : str = custom_convert_ldm_vae_checkpoint(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Dict = AutoencoderKL(**UpperCamelCase ) vae.load_state_dict(UpperCamelCase ) vae.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") _UpperCAmelCase : int = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
683
1
'''simple docstring''' from collections.abc import Iterable from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> Optional[int]: _UpperCamelCase : int = value _UpperCamelCase : Node | None = None # Added in order to delete a node easier _UpperCamelCase : Node | None = None _UpperCamelCase : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> List[Any]: _UpperCamelCase : str = root def __str__( self ) -> str: return str(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if new_children is not None: # reset its kids _UpperCamelCase : Union[str, Any] = node.parent if node.parent is not None: # reset its parent if self.is_right(_snake_case ): # If it is the right children _UpperCamelCase : str = new_children else: _UpperCamelCase : Any = new_children else: _UpperCamelCase : Any = new_children def _lowercase ( self , _snake_case ) -> bool: if node.parent and node.parent.right: return node == node.parent.right return False def _lowercase ( self ) -> bool: return self.root is None def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : List[Any] = Node(_snake_case ) # create a new Node if self.empty(): # if Tree is empty _UpperCamelCase : Optional[Any] = new_node # set its root else: # Tree is not empty _UpperCamelCase : int = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: _UpperCamelCase : Union[str, Any] = new_node # We insert the new node in a leaf break else: _UpperCamelCase : Union[str, Any] = parent_node.left else: if parent_node.right is None: _UpperCamelCase : Any = new_node break else: _UpperCamelCase : str = parent_node.right _UpperCamelCase : Any = parent_node def _lowercase ( self , *_snake_case ) -> None: for value in values: self.__insert(_snake_case ) def _lowercase ( self , _snake_case ) -> Node | None: if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: _UpperCamelCase : List[str] = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: _UpperCamelCase : Optional[Any] = node.left if value < node.value else node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: if self.root is None: return None _UpperCamelCase : Dict = self.root if not self.empty(): while node.right is not None: _UpperCamelCase : Tuple = node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: _UpperCamelCase : Optional[Any] = self.root if self.root is None: return None if not self.empty(): _UpperCamelCase : Optional[int] = self.root while node.left is not None: _UpperCamelCase : List[str] = node.left return node def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : str = self.search(_snake_case ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_snake_case , _snake_case ) elif node.left is None: # Has only right children self.__reassign_nodes(_snake_case , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_snake_case , node.left ) else: _UpperCamelCase : List[str] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore _UpperCamelCase : int = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _lowercase ( self , _snake_case ) -> Iterable: if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _lowercase ( self , _snake_case=None ) -> Any: if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if node: self.inorder(_snake_case , node.left ) arr.append(node.value ) self.inorder(_snake_case , node.right ) def _lowercase ( self , _snake_case , _snake_case ) -> int: _UpperCamelCase : list[int] = [] self.inorder(_snake_case , _snake_case ) # append all values to list using inorder traversal return arr[k - 1] def snake_case__ ( UpperCamelCase ) -> list[Node]: _UpperCamelCase : int = [] if curr_node is not None: _UpperCamelCase : Any = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def snake_case__ ( ) -> None: _UpperCamelCase : Any = (8, 3, 6, 1, 10, 14, 13, 4, 7) _UpperCamelCase : Tuple = BinarySearchTree() for i in testlist: t.insert(UpperCamelCase ) # Prints all the elements of the list in order traversal print(UpperCamelCase ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' ,t.get_max().value ) # type: ignore print('''Min Value: ''' ,t.get_min().value ) # type: ignore for i in testlist: t.remove(UpperCamelCase ) print(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
683
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = ['image_processor', 'tokenizer'] A__ : Dict = 'CLIPImageProcessor' A__ : str = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , _snake_case=None , _snake_case=None , **_snake_case ) -> List[Any]: _UpperCamelCase : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) _UpperCamelCase : Optional[Any] = kwargs.pop('''feature_extractor''' ) _UpperCamelCase : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_snake_case , _snake_case ) def __call__( self , _snake_case=None , _snake_case=None , _snake_case=None , **_snake_case ) -> Dict: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: _UpperCamelCase : List[str] = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: _UpperCamelCase : str = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: _UpperCamelCase : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Tuple: return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Any: return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def _lowercase ( self ) -> int: _UpperCamelCase : Optional[int] = self.tokenizer.model_input_names _UpperCamelCase : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Tuple: _UpperCamelCase : Optional[Any] = '''''' for i in table: res += inp[i - 1] return res def snake_case__ ( UpperCamelCase ) -> List[Any]: return data[1:] + data[0] def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> str: _UpperCamelCase : Dict = '''''' for i in range(len(UpperCamelCase ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Any = int('''0b''' + data[0] + data[-1] ,2 ) _UpperCamelCase : Any = int('''0b''' + data[1:3] ,2 ) return bin(s[row][col] )[2:] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: _UpperCamelCase : Tuple = message[:4] _UpperCamelCase : List[Any] = message[4:] _UpperCamelCase : List[Any] = apply_table(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : str = xor(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Any = apply_sbox(UpperCamelCase ,temp[:4] ) # noqa: E741 _UpperCamelCase : int = apply_sbox(UpperCamelCase ,temp[4:] ) _UpperCamelCase : str = '''0''' * (2 - len(UpperCamelCase )) + l # noqa: E741 _UpperCamelCase : Dict = '''0''' * (2 - len(UpperCamelCase )) + r _UpperCamelCase : Dict = apply_table(l + r ,UpperCamelCase ) _UpperCamelCase : List[str] = xor(UpperCamelCase ,UpperCamelCase ) return temp + right if __name__ == "__main__": _UpperCAmelCase : Union[str, Any] = input("""Enter 10 bit key: """) _UpperCAmelCase : Any = input("""Enter 8 bit message: """) _UpperCAmelCase : Tuple = [6, 3, 7, 4, 8, 5, 10, 9] _UpperCAmelCase : List[str] = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] _UpperCAmelCase : Dict = [2, 4, 3, 1] _UpperCAmelCase : Optional[int] = [2, 6, 3, 1, 4, 8, 5, 7] _UpperCAmelCase : Tuple = [4, 1, 3, 5, 7, 2, 8, 6] _UpperCAmelCase : List[Any] = [4, 1, 2, 3, 2, 3, 4, 1] _UpperCAmelCase : List[Any] = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] _UpperCAmelCase : Union[str, Any] = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation _UpperCAmelCase : List[str] = apply_table(key, paa_table) _UpperCAmelCase : Optional[Any] = temp[:5] _UpperCAmelCase : Union[str, Any] = temp[5:] _UpperCAmelCase : Dict = left_shift(left) _UpperCAmelCase : List[Any] = left_shift(right) _UpperCAmelCase : str = apply_table(left + right, pa_table) _UpperCAmelCase : Tuple = left_shift(left) _UpperCAmelCase : List[str] = left_shift(right) _UpperCAmelCase : Any = left_shift(left) _UpperCAmelCase : Union[str, Any] = left_shift(right) _UpperCAmelCase : Optional[Any] = apply_table(left + right, pa_table) # encryption _UpperCAmelCase : List[Any] = apply_table(message, IP) _UpperCAmelCase : List[Any] = function(expansion, sa, sa, keya, temp) _UpperCAmelCase : str = temp[4:] + temp[:4] _UpperCAmelCase : int = function(expansion, sa, sa, keya, temp) _UpperCAmelCase : Tuple = apply_table(temp, IP_inv) print("""Cipher text is:""", CT) # decryption _UpperCAmelCase : Dict = apply_table(CT, IP) _UpperCAmelCase : Optional[Any] = function(expansion, sa, sa, keya, temp) _UpperCAmelCase : Optional[Any] = temp[4:] + temp[:4] _UpperCAmelCase : List[Any] = function(expansion, sa, sa, keya, temp) _UpperCAmelCase : Union[str, Any] = apply_table(temp, IP_inv) print("""Plain text after decypting is:""", PT)
683
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _UpperCAmelCase : Union[str, Any] = (720, 1280) # Height, Width _UpperCAmelCase : str = (0.4, 0.6) # if height or width lower than this scale, drop it. _UpperCAmelCase : Optional[Any] = 1 / 100 _UpperCAmelCase : Optional[Any] = """""" _UpperCAmelCase : int = """""" _UpperCAmelCase : Union[str, Any] = """""" _UpperCAmelCase : List[Any] = 250 def snake_case__ ( ) -> None: _UpperCamelCase, _UpperCamelCase : List[Any] = get_dataset(UpperCamelCase ,UpperCamelCase ) for index in range(UpperCamelCase ): _UpperCamelCase : List[str] = random.sample(range(len(UpperCamelCase ) ) ,4 ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[str] = update_image_and_anno( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,filter_scale=UpperCamelCase ,) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _UpperCamelCase : List[str] = random_chars(32 ) _UpperCamelCase : List[str] = path.split(os.sep )[-1].rsplit('''.''' ,1 )[0] _UpperCamelCase : Any = f'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(f'''{file_root}.jpg''' ,UpperCamelCase ,[cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) _UpperCamelCase : Any = [] for anno in new_annos: _UpperCamelCase : List[Any] = anno[3] - anno[1] _UpperCamelCase : int = anno[4] - anno[2] _UpperCamelCase : int = anno[1] + width / 2 _UpperCamelCase : int = anno[2] + height / 2 _UpperCamelCase : Optional[Any] = f'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(UpperCamelCase ) with open(f'''{file_root}.txt''' ,'''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> tuple[list, list]: _UpperCamelCase : List[str] = [] _UpperCamelCase : Union[str, Any] = [] for label_file in glob.glob(os.path.join(UpperCamelCase ,'''*.txt''' ) ): _UpperCamelCase : int = label_file.split(os.sep )[-1].rsplit('''.''' ,1 )[0] with open(UpperCamelCase ) as in_file: _UpperCamelCase : Dict = in_file.readlines() _UpperCamelCase : Tuple = os.path.join(UpperCamelCase ,f'''{label_name}.jpg''' ) _UpperCamelCase : Tuple = [] for obj_list in obj_lists: _UpperCamelCase : List[Any] = obj_list.rstrip('''\n''' ).split(''' ''' ) _UpperCamelCase : Tuple = float(obj[1] ) - float(obj[3] ) / 2 _UpperCamelCase : Any = float(obj[2] ) - float(obj[4] ) / 2 _UpperCamelCase : Tuple = float(obj[1] ) + float(obj[3] ) / 2 _UpperCamelCase : List[Any] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(UpperCamelCase ) labels.append(UpperCamelCase ) return img_paths, labels def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = 0.0 ,) -> tuple[list, list, str]: _UpperCamelCase : Optional[int] = np.zeros([output_size[0], output_size[1], 3] ,dtype=np.uinta ) _UpperCamelCase : str = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = int(scale_x * output_size[1] ) _UpperCamelCase : Dict = int(scale_y * output_size[0] ) _UpperCamelCase : int = [] _UpperCamelCase : Union[str, Any] = [] for i, index in enumerate(UpperCamelCase ): _UpperCamelCase : Optional[int] = all_img_list[index] path_list.append(UpperCamelCase ) _UpperCamelCase : str = all_annos[index] _UpperCamelCase : Tuple = cva.imread(UpperCamelCase ) if i == 0: # top-left _UpperCamelCase : Any = cva.resize(UpperCamelCase ,(divid_point_x, divid_point_y) ) _UpperCamelCase : Any = img for bbox in img_annos: _UpperCamelCase : List[Any] = bbox[1] * scale_x _UpperCamelCase : Dict = bbox[2] * scale_y _UpperCamelCase : Any = bbox[3] * scale_x _UpperCamelCase : Any = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _UpperCamelCase : Union[str, Any] = cva.resize(UpperCamelCase ,(output_size[1] - divid_point_x, divid_point_y) ) _UpperCamelCase : List[Any] = img for bbox in img_annos: _UpperCamelCase : Any = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Optional[Any] = bbox[2] * scale_y _UpperCamelCase : Any = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : Optional[int] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _UpperCamelCase : Dict = cva.resize(UpperCamelCase ,(divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : List[str] = img for bbox in img_annos: _UpperCamelCase : int = bbox[1] * scale_x _UpperCamelCase : Optional[Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : int = bbox[3] * scale_x _UpperCamelCase : Any = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _UpperCamelCase : Dict = cva.resize( UpperCamelCase ,(output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : Union[str, Any] = img for bbox in img_annos: _UpperCamelCase : Optional[int] = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Union[str, Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : List[str] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _UpperCamelCase : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def snake_case__ ( UpperCamelCase ) -> str: assert number_char > 1, "The number of character should greater than 1" _UpperCamelCase : Tuple = ascii_lowercase + digits return "".join(random.choice(UpperCamelCase ) for _ in range(UpperCamelCase ) ) if __name__ == "__main__": main() print("""DONE ✅""")
683
1
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) _UpperCAmelCase : str = ["""model.decoder.embed_positions.weights"""] def snake_case__ ( UpperCamelCase ) -> str: if "emb" in name: _UpperCamelCase : List[str] = name.replace('''emb''' ,'''model.decoder.embed_tokens''' ) if "transformer" in name: _UpperCamelCase : Optional[Any] = name.replace('''transformer''' ,'''model.decoder''' ) if "cross_attention" in name: _UpperCamelCase : Optional[Any] = name.replace('''cross_attention''' ,'''encoder_attn''' ) if "linear1" in name: _UpperCamelCase : Optional[int] = name.replace('''linear1''' ,'''fc1''' ) if "linear2" in name: _UpperCamelCase : str = name.replace('''linear2''' ,'''fc2''' ) if "norm1" in name: _UpperCamelCase : Optional[Any] = name.replace('''norm1''' ,'''self_attn_layer_norm''' ) if "norm_cross" in name: _UpperCamelCase : List[str] = name.replace('''norm_cross''' ,'''encoder_attn_layer_norm''' ) if "norm2" in name: _UpperCamelCase : Any = name.replace('''norm2''' ,'''final_layer_norm''' ) if "out_norm" in name: _UpperCamelCase : str = name.replace('''out_norm''' ,'''model.decoder.layer_norm''' ) if "linears" in name: _UpperCamelCase : Dict = name.replace('''linears''' ,'''lm_heads''' ) if "condition_provider.conditioners.description.output_proj" in name: _UpperCamelCase : int = name.replace('''condition_provider.conditioners.description.output_proj''' ,'''enc_to_dec_proj''' ) return name def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Tuple[Dict, Dict]: _UpperCamelCase : Optional[int] = list(state_dict.keys() ) _UpperCamelCase : Union[str, Any] = {} for key in keys: _UpperCamelCase : Union[str, Any] = state_dict.pop(UpperCamelCase ) _UpperCamelCase : Optional[Any] = rename_keys(UpperCamelCase ) if "in_proj_weight" in key: # split fused qkv proj _UpperCamelCase : Optional[int] = val[:hidden_size, :] _UpperCamelCase : int = val[hidden_size : 2 * hidden_size, :] _UpperCamelCase : Optional[int] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: _UpperCamelCase : Dict = val else: _UpperCamelCase : List[str] = val return state_dict, enc_dec_proj_state_dict def snake_case__ ( UpperCamelCase ) -> MusicgenDecoderConfig: if checkpoint == "small": # default config values _UpperCamelCase : List[str] = 10_24 _UpperCamelCase : List[Any] = 24 _UpperCamelCase : int = 16 elif checkpoint == "medium": _UpperCamelCase : str = 15_36 _UpperCamelCase : Union[str, Any] = 48 _UpperCamelCase : Dict = 24 elif checkpoint == "large": _UpperCamelCase : Tuple = 20_48 _UpperCamelCase : Optional[Any] = 48 _UpperCamelCase : Union[str, Any] = 32 else: raise ValueError(f'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) _UpperCamelCase : Optional[Any] = MusicgenDecoderConfig( hidden_size=UpperCamelCase ,ffn_dim=hidden_size * 4 ,num_hidden_layers=UpperCamelCase ,num_attention_heads=UpperCamelCase ,) return config @torch.no_grad() def snake_case__ ( UpperCamelCase ,UpperCamelCase=None ,UpperCamelCase=None ,UpperCamelCase="cpu" ) -> List[str]: _UpperCamelCase : Tuple = MusicGen.get_pretrained(UpperCamelCase ,device=UpperCamelCase ) _UpperCamelCase : Optional[int] = decoder_config_from_checkpoint(UpperCamelCase ) _UpperCamelCase : List[Any] = fairseq_model.lm.state_dict() _UpperCamelCase, _UpperCamelCase : int = rename_state_dict( UpperCamelCase ,hidden_size=decoder_config.hidden_size ) _UpperCamelCase : List[str] = TaEncoderModel.from_pretrained('''t5-base''' ) _UpperCamelCase : List[str] = EncodecModel.from_pretrained('''facebook/encodec_32khz''' ) _UpperCamelCase : Any = MusicgenForCausalLM(UpperCamelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection _UpperCamelCase, _UpperCamelCase : List[str] = decoder.load_state_dict(UpperCamelCase ,strict=UpperCamelCase ) for key in missing_keys.copy(): if key.startswith(('''text_encoder''', '''audio_encoder''') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Missing key(s) in state_dict: {missing_keys}''' ) if len(UpperCamelCase ) > 0: raise ValueError(f'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model _UpperCamelCase : Dict = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase ,audio_encoder=UpperCamelCase ,decoder=UpperCamelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase ) # check we can do a forward pass _UpperCamelCase : str = torch.arange(0 ,8 ,dtype=torch.long ).reshape(2 ,-1 ) _UpperCamelCase : int = input_ids.reshape(2 * 4 ,-1 ) with torch.no_grad(): _UpperCamelCase : str = model(input_ids=UpperCamelCase ,decoder_input_ids=UpperCamelCase ).logits if logits.shape != (8, 1, 20_48): raise ValueError('''Incorrect shape for logits''' ) # now construct the processor _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained('''t5-base''' ) _UpperCamelCase : Optional[int] = AutoFeatureExtractor.from_pretrained('''facebook/encodec_32khz''' ,padding_side='''left''' ) _UpperCamelCase : Tuple = MusicgenProcessor(feature_extractor=UpperCamelCase ,tokenizer=UpperCamelCase ) # set the appropriate bos/pad token ids _UpperCamelCase : Optional[Any] = 20_48 _UpperCamelCase : Tuple = 20_48 # set other default generation config params _UpperCamelCase : Tuple = int(30 * audio_encoder.config.frame_rate ) _UpperCamelCase : int = True _UpperCamelCase : Dict = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) logger.info(f'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if repo_id: logger.info(f'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(UpperCamelCase ) processor.push_to_hub(UpperCamelCase ) if __name__ == "__main__": _UpperCAmelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint""", default="""small""", type=str, help="""Checkpoint size of the MusicGen model you'd like to convert. Can be one of: `['small', 'medium', 'large']`.""", ) parser.add_argument( """--pytorch_dump_folder""", required=True, default=None, type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) parser.add_argument( """--device""", default="""cpu""", type=str, help="""Torch device to run the conversion, either cpu or cuda.""" ) _UpperCAmelCase : Dict = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
683
'''simple docstring''' from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class UpperCAmelCase ( a_ ): """simple docstring""" @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Union[str, Any] = EncoderDecoderModel.from_encoder_decoder_pretrained('''prajjwal1/bert-tiny''' , '''prajjwal1/bert-tiny''' ) _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained('''bert-base-uncased''' ) _UpperCamelCase : Optional[Any] = bertabert.config.encoder.vocab_size _UpperCamelCase : List[str] = tokenizer.sep_token_id _UpperCamelCase : List[str] = tokenizer.cls_token_id _UpperCamelCase : Optional[Any] = 128 _UpperCamelCase : int = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''train[:1%]''' ) _UpperCamelCase : Dict = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''validation[:1%]''' ) _UpperCamelCase : Dict = train_dataset.select(range(32 ) ) _UpperCamelCase : Tuple = val_dataset.select(range(16 ) ) _UpperCamelCase : Union[str, Any] = 4 def _map_to_encoder_decoder_inputs(_snake_case ): # Tokenizer will automatically set [BOS] <text> [EOS] _UpperCamelCase : Optional[Any] = tokenizer(batch['''article'''] , padding='''max_length''' , truncation=_snake_case , max_length=512 ) _UpperCamelCase : Optional[int] = tokenizer(batch['''highlights'''] , padding='''max_length''' , truncation=_snake_case , max_length=128 ) _UpperCamelCase : str = inputs.input_ids _UpperCamelCase : Union[str, Any] = inputs.attention_mask _UpperCamelCase : str = outputs.input_ids _UpperCamelCase : str = outputs.input_ids.copy() _UpperCamelCase : Tuple = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['''labels'''] ] _UpperCamelCase : Union[str, Any] = outputs.attention_mask assert all(len(_snake_case ) == 512 for x in inputs.input_ids ) assert all(len(_snake_case ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_snake_case ): _UpperCamelCase : Dict = pred.label_ids _UpperCamelCase : Optional[int] = pred.predictions # all unnecessary tokens are removed _UpperCamelCase : Any = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : Dict = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_snake_case ) )] ) / len(_snake_case ) return {"accuracy": accuracy} # map train dataset _UpperCamelCase : Optional[Any] = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) train_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) # same for validation dataset _UpperCamelCase : List[Any] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) val_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) _UpperCamelCase : Optional[int] = self.get_auto_remove_tmp_dir() _UpperCamelCase : Union[str, Any] = SeqaSeqTrainingArguments( output_dir=_snake_case , per_device_train_batch_size=_snake_case , per_device_eval_batch_size=_snake_case , predict_with_generate=_snake_case , evaluation_strategy='''steps''' , do_train=_snake_case , do_eval=_snake_case , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _UpperCamelCase : Optional[int] = SeqaSeqTrainer( model=_snake_case , args=_snake_case , compute_metrics=_compute_metrics , train_dataset=_snake_case , eval_dataset=_snake_case , tokenizer=_snake_case , ) # start training trainer.train()
683
1
'''simple docstring''' from collections import defaultdict class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case , _snake_case ) -> Optional[Any]: _UpperCamelCase : Optional[Any] = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 _UpperCamelCase : Optional[Any] = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(_snake_case ) ) ] _UpperCamelCase : List[str] = defaultdict(_snake_case ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 _UpperCamelCase : Optional[int] = (1 << len(_snake_case )) - 1 def _lowercase ( self , _snake_case , _snake_case ) -> Optional[Any]: # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement _UpperCamelCase : Optional[Any] = self.count_ways_until(_snake_case , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. _UpperCamelCase : Dict = total_ways_util return self.dp[mask][task_no] def _lowercase ( self , _snake_case ) -> List[Any]: # Store the list of persons for each task for i in range(len(_snake_case ) ): for j in task_performed[i]: self.task[j].append(_snake_case ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": _UpperCAmelCase : Tuple = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. _UpperCAmelCase : Optional[int] = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
683
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def snake_case__ ( UpperCamelCase=None ) -> Optional[int]: if subparsers is not None: _UpperCamelCase : Dict = subparsers.add_parser('''env''' ) else: _UpperCamelCase : Tuple = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''' ,default=UpperCamelCase ,help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase ) return parser def snake_case__ ( UpperCamelCase ) -> Any: _UpperCamelCase : int = torch.__version__ _UpperCamelCase : int = torch.cuda.is_available() _UpperCamelCase : List[str] = is_xpu_available() _UpperCamelCase : Dict = is_npu_available() _UpperCamelCase : Optional[Any] = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(UpperCamelCase ): _UpperCamelCase : List[str] = load_config_from_file(args.config_file ).to_dict() _UpperCamelCase : List[Any] = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': f'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(UpperCamelCase ), '''PyTorch NPU available''': str(UpperCamelCase ), '''System RAM''': f'''{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB''', } if pt_cuda_available: _UpperCamelCase : int = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([f'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) _UpperCamelCase : Union[str, Any] = ( '''\n'''.join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(UpperCamelCase ,UpperCamelCase ) else f'''\t{accelerate_config}''' ) print(UpperCamelCase ) _UpperCamelCase : str = accelerate_config return info def snake_case__ ( ) -> int: _UpperCamelCase : str = env_command_parser() _UpperCamelCase : Any = parser.parse_args() env_command(UpperCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
683
1
'''simple docstring''' import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _UpperCAmelCase : Optional[int] = 2 class UpperCAmelCase : """simple docstring""" def __init__( self , *, # begin keyword-only arguments _snake_case="<s>" , _snake_case="<pad>" , _snake_case="</s>" , _snake_case="<unk>" , _snake_case=None , ) -> Union[str, Any]: _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Union[str, Any] = bos, unk, pad, eos _UpperCamelCase : Optional[Any] = [] _UpperCamelCase : Tuple = [] _UpperCamelCase : str = {} _UpperCamelCase : List[str] = self.add_symbol(_snake_case ) _UpperCamelCase : Any = self.add_symbol(_snake_case ) _UpperCamelCase : Tuple = self.add_symbol(_snake_case ) _UpperCamelCase : List[Any] = self.add_symbol(_snake_case ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(_snake_case ) _UpperCamelCase : str = len(self.symbols ) def __eq__( self , _snake_case ) -> List[Any]: return self.indices == other.indices def __getitem__( self , _snake_case ) -> Dict: if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ) -> Dict: return len(self.symbols ) def __contains__( self , _snake_case ) -> str: return sym in self.indices @classmethod def _lowercase ( cls , _snake_case ) -> str: _UpperCamelCase : int = cls() d.add_from_file(_snake_case ) return d def _lowercase ( self , _snake_case , _snake_case=1 , _snake_case=False ) -> List[str]: if word in self.indices and not overwrite: _UpperCamelCase : int = self.indices[word] _UpperCamelCase : List[Any] = self.count[idx] + n return idx else: _UpperCamelCase : List[str] = len(self.symbols ) _UpperCamelCase : str = idx self.symbols.append(_snake_case ) self.count.append(_snake_case ) return idx def _lowercase ( self , _snake_case ) -> Union[str, Any]: return 0 def _lowercase ( self , _snake_case ) -> Dict: if isinstance(_snake_case , _snake_case ): try: with open(_snake_case , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(_snake_case ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(_snake_case ) ) return _UpperCamelCase : str = f.readlines() _UpperCamelCase : Tuple = self._load_meta(_snake_case ) for line in lines[indices_start_line:]: try: _UpperCamelCase, _UpperCamelCase : Optional[int] = line.rstrip().rsplit(''' ''' , 1 ) if field == "#fairseq:overwrite": _UpperCamelCase : Optional[int] = True _UpperCamelCase, _UpperCamelCase : Optional[int] = line.rsplit(''' ''' , 1 ) else: _UpperCamelCase : Tuple = False _UpperCamelCase : List[str] = int(_snake_case ) _UpperCamelCase : Union[str, Any] = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(_snake_case ) ) self.add_symbol(_snake_case , n=_snake_case , overwrite=_snake_case ) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''' ) def snake_case__ ( UpperCamelCase ) -> List[Any]: # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} _UpperCamelCase : List[Any] = dict((re.sub(r'''@@$''' ,'''''' ,UpperCamelCase ), v) if k.endswith('''@@''' ) else (re.sub(r'''$''' ,'''</w>''' ,UpperCamelCase ), v) for k, v in d.items() ) _UpperCamelCase : List[Any] = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[f'''{k}</w>'''] _UpperCamelCase : Optional[Any] = d[k] # restore return da def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Tuple: # prep if not os.path.exists(UpperCamelCase ): raise ValueError(f'''path {biogpt_checkpoint_path} does not exist!''' ) os.makedirs(UpperCamelCase ,exist_ok=UpperCamelCase ) print(f'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models _UpperCamelCase : Union[str, Any] = os.path.join(UpperCamelCase ,'''checkpoint.pt''' ) if not os.path.isfile(UpperCamelCase ): raise ValueError(f'''path to the file {checkpoint_file} does not exist!''' ) _UpperCamelCase : Union[str, Any] = torch.load(UpperCamelCase ,map_location='''cpu''' ) _UpperCamelCase : List[str] = chkpt['''cfg''']['''model'''] # dicts _UpperCamelCase : Union[str, Any] = os.path.join(UpperCamelCase ,'''dict.txt''' ) if not os.path.isfile(UpperCamelCase ): raise ValueError(f'''path to the file {dict_file} does not exist!''' ) _UpperCamelCase : Optional[int] = Dictionary.load(UpperCamelCase ) _UpperCamelCase : List[str] = rewrite_dict_keys(src_dict.indices ) _UpperCamelCase : Any = len(UpperCamelCase ) _UpperCamelCase : List[Any] = os.path.join(UpperCamelCase ,VOCAB_FILES_NAMES['''vocab_file'''] ) print(f'''Generating {src_vocab_file} of {src_vocab_size} records''' ) with open(UpperCamelCase ,'''w''' ,encoding='''utf-8''' ) as f: f.write(json.dumps(UpperCamelCase ,ensure_ascii=UpperCamelCase ,indent=UpperCamelCase ) ) # merges_file (bpecodes) _UpperCamelCase : Tuple = os.path.join(UpperCamelCase ,'''bpecodes''' ) if not os.path.isfile(UpperCamelCase ): raise ValueError(f'''path to the file {bpecodes_file} does not exist!''' ) _UpperCamelCase : int = os.path.join(UpperCamelCase ,VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(UpperCamelCase ,UpperCamelCase ) # model config _UpperCamelCase : str = os.path.join(UpperCamelCase ,'''config.json''' ) _UpperCamelCase : List[Any] = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.02, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1e-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(f'''Generating {biogpt_model_config_file}''' ) with open(UpperCamelCase ,'''w''' ,encoding='''utf-8''' ) as f: f.write(json.dumps(UpperCamelCase ,ensure_ascii=UpperCamelCase ,indent=UpperCamelCase ) ) # tokenizer config _UpperCamelCase : Optional[Any] = os.path.join(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : int = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 10_24, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(f'''Generating {biogpt_tokenizer_config_file}''' ) with open(UpperCamelCase ,'''w''' ,encoding='''utf-8''' ) as f: f.write(json.dumps(UpperCamelCase ,ensure_ascii=UpperCamelCase ,indent=UpperCamelCase ) ) # model _UpperCamelCase : Union[str, Any] = chkpt['''model'''] # remove unneeded keys _UpperCamelCase : List[str] = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[Any] = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): _UpperCamelCase : List[str] = model_state_dict.pop(UpperCamelCase ) else: _UpperCamelCase : List[str] = model_state_dict.pop(UpperCamelCase ) _UpperCamelCase : Any = BioGptConfig.from_pretrained(UpperCamelCase ) _UpperCamelCase : List[str] = BioGptForCausalLM(UpperCamelCase ) # check that it loads ok model_new.load_state_dict(UpperCamelCase ) # save _UpperCamelCase : int = os.path.join(UpperCamelCase ,UpperCamelCase ) print(f'''Generating {pytorch_weights_dump_path}''' ) torch.save(UpperCamelCase ,UpperCamelCase ) print('''Conversion is done!''' ) if __name__ == "__main__": _UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--biogpt_checkpoint_path""", default=None, type=str, required=True, help=( """Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,""" """ bpecodes, etc.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _UpperCAmelCase : List[Any] = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
683
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : List[Any] = logging.get_logger(__name__) def snake_case__ ( UpperCamelCase ) -> Tuple: _UpperCamelCase : str = '''huggingface/label-files''' _UpperCamelCase : Optional[Any] = '''imagenet-1k-id2label.json''' _UpperCamelCase : Optional[int] = json.load(open(hf_hub_download(UpperCamelCase ,UpperCamelCase ,repo_type='''dataset''' ) ,'''r''' ) ) _UpperCamelCase : Optional[int] = {int(UpperCamelCase ): v for k, v in idalabel.items()} _UpperCamelCase : Dict = {v: k for k, v in idalabel.items()} _UpperCamelCase : Optional[Any] = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" _UpperCamelCase : Union[str, Any] = BitConfig( conv_layer=UpperCamelCase ,num_labels=10_00 ,idalabel=UpperCamelCase ,labelaid=UpperCamelCase ,) return config def snake_case__ ( UpperCamelCase ) -> str: if "stem.conv" in name: _UpperCamelCase : Any = name.replace('''stem.conv''' ,'''bit.embedder.convolution''' ) if "blocks" in name: _UpperCamelCase : Union[str, Any] = name.replace('''blocks''' ,'''layers''' ) if "head.fc" in name: _UpperCamelCase : Optional[Any] = name.replace('''head.fc''' ,'''classifier.1''' ) if name.startswith('''norm''' ): _UpperCamelCase : Any = '''bit.''' + name if "bit" not in name and "classifier" not in name: _UpperCamelCase : List[Any] = '''bit.encoder.''' + name return name def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase : List[str] = Image.open(requests.get(UpperCamelCase ,stream=UpperCamelCase ).raw ) return im @torch.no_grad() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[Any]: _UpperCamelCase : str = get_config(UpperCamelCase ) # load original model from timm _UpperCamelCase : int = create_model(UpperCamelCase ,pretrained=UpperCamelCase ) timm_model.eval() # load state_dict of original model _UpperCamelCase : int = timm_model.state_dict() for key in state_dict.copy().keys(): _UpperCamelCase : int = state_dict.pop(UpperCamelCase ) _UpperCamelCase : Any = val.squeeze() if '''head''' in key else val # load HuggingFace model _UpperCamelCase : List[str] = BitForImageClassification(UpperCamelCase ) model.eval() model.load_state_dict(UpperCamelCase ) # create image processor _UpperCamelCase : Optional[int] = create_transform(**resolve_data_config({} ,model=UpperCamelCase ) ) _UpperCamelCase : Any = transform.transforms _UpperCamelCase : List[str] = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } _UpperCamelCase : List[str] = BitImageProcessor( do_resize=UpperCamelCase ,size={'''shortest_edge''': timm_transforms[0].size} ,resample=pillow_resamplings[timm_transforms[0].interpolation.value] ,do_center_crop=UpperCamelCase ,crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} ,do_normalize=UpperCamelCase ,image_mean=timm_transforms[-1].mean.tolist() ,image_std=timm_transforms[-1].std.tolist() ,) _UpperCamelCase : str = prepare_img() _UpperCamelCase : Dict = transform(UpperCamelCase ).unsqueeze(0 ) _UpperCamelCase : Dict = processor(UpperCamelCase ,return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(UpperCamelCase ,UpperCamelCase ) # verify logits with torch.no_grad(): _UpperCamelCase : Optional[int] = model(UpperCamelCase ) _UpperCamelCase : Optional[int] = outputs.logits print('''Logits:''' ,logits[0, :3] ) print('''Predicted class:''' ,model.config.idalabel[logits.argmax(-1 ).item()] ) _UpperCamelCase : List[Any] = timm_model(UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase ,outputs.logits ,atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": _UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) _UpperCAmelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
683
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case , _snake_case=13 , _snake_case=7 , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=99 , _snake_case=[1, 1, 2] , _snake_case=1 , _snake_case=32 , _snake_case=4 , _snake_case=8 , _snake_case=37 , _snake_case="gelu_new" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=0.0 , _snake_case=512 , _snake_case=3 , _snake_case=0.02 , _snake_case=3 , _snake_case=4 , _snake_case=None , _snake_case=False , ) -> str: _UpperCamelCase : Tuple = parent _UpperCamelCase : int = batch_size _UpperCamelCase : Optional[Any] = seq_length _UpperCamelCase : Union[str, Any] = is_training _UpperCamelCase : List[str] = use_input_mask _UpperCamelCase : str = use_token_type_ids _UpperCamelCase : Union[str, Any] = use_labels _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Optional[int] = block_sizes _UpperCamelCase : Dict = num_decoder_layers _UpperCamelCase : List[Any] = d_model _UpperCamelCase : int = n_head _UpperCamelCase : Optional[Any] = d_head _UpperCamelCase : Optional[int] = d_inner _UpperCamelCase : int = hidden_act _UpperCamelCase : Dict = hidden_dropout _UpperCamelCase : int = attention_dropout _UpperCamelCase : Tuple = activation_dropout _UpperCamelCase : Any = max_position_embeddings _UpperCamelCase : str = type_vocab_size _UpperCamelCase : Optional[int] = 2 _UpperCamelCase : List[Any] = num_labels _UpperCamelCase : Optional[Any] = num_choices _UpperCamelCase : Union[str, Any] = scope _UpperCamelCase : Tuple = initializer_std # Used in the tests to check the size of the first attention layer _UpperCamelCase : List[str] = n_head # Used in the tests to check the size of the first hidden state _UpperCamelCase : Any = self.d_model # Used in the tests to check the number of output hidden states/attentions _UpperCamelCase : List[Any] = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: _UpperCamelCase : Optional[Any] = self.num_hidden_layers + 2 def _lowercase ( self ) -> int: _UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase : Optional[int] = None if self.use_input_mask: _UpperCamelCase : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = None if self.use_token_type_ids: _UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase : int = None _UpperCamelCase : Optional[int] = None _UpperCamelCase : Optional[int] = None if self.use_labels: _UpperCamelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) _UpperCamelCase : Optional[Any] = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Optional[int]: _UpperCamelCase : int = TFFunnelModel(config=_snake_case ) _UpperCamelCase : Union[str, Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : Optional[Any] = model(_snake_case ) _UpperCamelCase : List[str] = [input_ids, input_mask] _UpperCamelCase : Dict = model(_snake_case ) _UpperCamelCase : str = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) _UpperCamelCase : Dict = False _UpperCamelCase : Union[str, Any] = TFFunnelModel(config=_snake_case ) _UpperCamelCase : Tuple = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) _UpperCamelCase : Tuple = False _UpperCamelCase : Any = TFFunnelModel(config=_snake_case ) _UpperCamelCase : Dict = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Tuple: _UpperCamelCase : Union[str, Any] = TFFunnelBaseModel(config=_snake_case ) _UpperCamelCase : Optional[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : int = model(_snake_case ) _UpperCamelCase : Optional[int] = [input_ids, input_mask] _UpperCamelCase : List[Any] = model(_snake_case ) _UpperCamelCase : Any = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) _UpperCamelCase : int = False _UpperCamelCase : Dict = TFFunnelBaseModel(config=_snake_case ) _UpperCamelCase : Tuple = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) _UpperCamelCase : Optional[int] = False _UpperCamelCase : str = TFFunnelBaseModel(config=_snake_case ) _UpperCamelCase : str = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> List[Any]: _UpperCamelCase : Union[str, Any] = TFFunnelForPreTraining(config=_snake_case ) _UpperCamelCase : Union[str, Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : Optional[Any] = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> str: _UpperCamelCase : List[Any] = TFFunnelForMaskedLM(config=_snake_case ) _UpperCamelCase : Optional[int] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : Optional[int] = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Optional[Any]: _UpperCamelCase : Optional[Any] = self.num_labels _UpperCamelCase : Dict = TFFunnelForSequenceClassification(config=_snake_case ) _UpperCamelCase : Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : Union[str, Any] = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Dict: _UpperCamelCase : Dict = self.num_choices _UpperCamelCase : Tuple = TFFunnelForMultipleChoice(config=_snake_case ) _UpperCamelCase : Dict = tf.tile(tf.expand_dims(_snake_case , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(_snake_case , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase : Any = tf.tile(tf.expand_dims(_snake_case , 1 ) , (1, self.num_choices, 1) ) _UpperCamelCase : Optional[int] = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } _UpperCamelCase : Dict = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Any: _UpperCamelCase : Tuple = self.num_labels _UpperCamelCase : Any = TFFunnelForTokenClassification(config=_snake_case ) _UpperCamelCase : Dict = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : int = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ) -> Dict: _UpperCamelCase : Optional[Any] = TFFunnelForQuestionAnswering(config=_snake_case ) _UpperCamelCase : Optional[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _UpperCamelCase : Optional[Any] = model(_snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self ) -> List[str]: _UpperCamelCase : Dict = self.prepare_config_and_inputs() ( ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ( _UpperCamelCase ), ) : Dict = config_and_inputs _UpperCamelCase : List[Any] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase ( a_ , a_ , unittest.TestCase ): """simple docstring""" A__ : Tuple = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) A__ : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) A__ : Tuple = False A__ : Tuple = False def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = TFFunnelModelTester(self ) _UpperCamelCase : Any = ConfigTester(self , config_class=_snake_case ) def _lowercase ( self ) -> Any: self.config_tester.run_common_tests() def _lowercase ( self ) -> str: _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_snake_case ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_snake_case ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_snake_case ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_snake_case ) @require_tf class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Dict = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) A__ : Optional[int] = False A__ : Any = False def _lowercase ( self ) -> List[Any]: _UpperCamelCase : List[str] = TFFunnelModelTester(self , base=_snake_case ) _UpperCamelCase : str = ConfigTester(self , config_class=_snake_case ) def _lowercase ( self ) -> List[Any]: self.config_tester.run_common_tests() def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*_snake_case ) def _lowercase ( self ) -> List[str]: _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_snake_case ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_snake_case )
683
'''simple docstring''' _UpperCAmelCase : Any = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100000)] def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : Any = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_00_00] number //= 10_00_00 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _UpperCAmelCase : list[bool | None] = [None] * 10000000 _UpperCAmelCase : str = True _UpperCAmelCase : Tuple = False def snake_case__ ( UpperCamelCase ) -> bool: if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _UpperCamelCase : List[str] = chain(next_number(UpperCamelCase ) ) _UpperCamelCase : Tuple = number_chain while number < 10_00_00_00: _UpperCamelCase : int = number_chain number *= 10 return number_chain def snake_case__ ( UpperCamelCase = 10_00_00_00 ) -> int: for i in range(1 ,UpperCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
683
1
'''simple docstring''' from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class UpperCAmelCase ( a_ ): """simple docstring""" @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Union[str, Any] = EncoderDecoderModel.from_encoder_decoder_pretrained('''prajjwal1/bert-tiny''' , '''prajjwal1/bert-tiny''' ) _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained('''bert-base-uncased''' ) _UpperCamelCase : Optional[Any] = bertabert.config.encoder.vocab_size _UpperCamelCase : List[str] = tokenizer.sep_token_id _UpperCamelCase : List[str] = tokenizer.cls_token_id _UpperCamelCase : Optional[Any] = 128 _UpperCamelCase : int = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''train[:1%]''' ) _UpperCamelCase : Dict = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''validation[:1%]''' ) _UpperCamelCase : Dict = train_dataset.select(range(32 ) ) _UpperCamelCase : Tuple = val_dataset.select(range(16 ) ) _UpperCamelCase : Union[str, Any] = 4 def _map_to_encoder_decoder_inputs(_snake_case ): # Tokenizer will automatically set [BOS] <text> [EOS] _UpperCamelCase : Optional[Any] = tokenizer(batch['''article'''] , padding='''max_length''' , truncation=_snake_case , max_length=512 ) _UpperCamelCase : Optional[int] = tokenizer(batch['''highlights'''] , padding='''max_length''' , truncation=_snake_case , max_length=128 ) _UpperCamelCase : str = inputs.input_ids _UpperCamelCase : Union[str, Any] = inputs.attention_mask _UpperCamelCase : str = outputs.input_ids _UpperCamelCase : str = outputs.input_ids.copy() _UpperCamelCase : Tuple = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['''labels'''] ] _UpperCamelCase : Union[str, Any] = outputs.attention_mask assert all(len(_snake_case ) == 512 for x in inputs.input_ids ) assert all(len(_snake_case ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_snake_case ): _UpperCamelCase : Dict = pred.label_ids _UpperCamelCase : Optional[int] = pred.predictions # all unnecessary tokens are removed _UpperCamelCase : Any = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : Dict = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_snake_case ) )] ) / len(_snake_case ) return {"accuracy": accuracy} # map train dataset _UpperCamelCase : Optional[Any] = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) train_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) # same for validation dataset _UpperCamelCase : List[Any] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) val_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) _UpperCamelCase : Optional[int] = self.get_auto_remove_tmp_dir() _UpperCamelCase : Union[str, Any] = SeqaSeqTrainingArguments( output_dir=_snake_case , per_device_train_batch_size=_snake_case , per_device_eval_batch_size=_snake_case , predict_with_generate=_snake_case , evaluation_strategy='''steps''' , do_train=_snake_case , do_eval=_snake_case , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _UpperCamelCase : Optional[int] = SeqaSeqTrainer( model=_snake_case , args=_snake_case , compute_metrics=_compute_metrics , train_dataset=_snake_case , eval_dataset=_snake_case , tokenizer=_snake_case , ) # start training trainer.train()
683
'''simple docstring''' _UpperCAmelCase : str = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : List[str] = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> str: assert len(str(UpperCamelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _UpperCamelCase : Any = year // 1_00 _UpperCamelCase : List[Any] = (5 * (century % 4) + 2) % 7 _UpperCamelCase : Tuple = year % 1_00 _UpperCamelCase : Optional[int] = centurian % 12 _UpperCamelCase : Tuple = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _UpperCamelCase : List[Any] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) _UpperCamelCase : Optional[int] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
683
1
'''simple docstring''' import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _snake_case , _snake_case=13 , _snake_case=7 , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=99 , _snake_case=32 , _snake_case=5 , _snake_case=4 , _snake_case=37 , _snake_case="gelu" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=16 , _snake_case=2 , _snake_case=0.02 , _snake_case=4 , ) -> Tuple: _UpperCamelCase : List[Any] = parent _UpperCamelCase : List[Any] = batch_size _UpperCamelCase : Any = seq_length _UpperCamelCase : str = is_training _UpperCamelCase : Tuple = use_attention_mask _UpperCamelCase : Optional[Any] = use_token_type_ids _UpperCamelCase : Optional[Any] = use_labels _UpperCamelCase : Tuple = vocab_size _UpperCamelCase : Optional[int] = hidden_size _UpperCamelCase : Any = num_hidden_layers _UpperCamelCase : Optional[Any] = num_attention_heads _UpperCamelCase : List[Any] = intermediate_size _UpperCamelCase : List[str] = hidden_act _UpperCamelCase : Tuple = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : Tuple = max_position_embeddings _UpperCamelCase : Optional[Any] = type_vocab_size _UpperCamelCase : Tuple = type_sequence_label_size _UpperCamelCase : Any = initializer_range _UpperCamelCase : List[Any] = num_choices def _lowercase ( self ) -> Tuple: _UpperCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase : Tuple = None if self.use_attention_mask: _UpperCamelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = None if self.use_token_type_ids: _UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase : Optional[Any] = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_snake_case , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _lowercase ( self ) -> Optional[Any]: _UpperCamelCase : Optional[Any] = self.prepare_config_and_inputs() _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : int = config_and_inputs _UpperCamelCase : Union[str, Any] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict def _lowercase ( self ) -> List[str]: _UpperCamelCase : Union[str, Any] = self.prepare_config_and_inputs() _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Any = True _UpperCamelCase : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _UpperCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : List[str] = True A__ : Union[str, Any] = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Union[str, Any] = FlaxBertModelTester(self ) @slow def _lowercase ( self ) -> Dict: # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. _UpperCamelCase : Optional[Any] = FlaxBertModel.from_pretrained('''bert-base-cased''' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_snake_case )
683
'''simple docstring''' import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : """simple docstring""" @staticmethod def _lowercase ( *_snake_case , **_snake_case ) -> str: pass @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" A__ : Tuple = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Optional[Any]: _UpperCamelCase : int = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Any = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def _lowercase ( self , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : int = vqa_pipeline(_snake_case , top_k=1 ) self.assertEqual( _snake_case , [ [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], ] , ) @require_torch def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[int] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question='''How many cats are there?''' , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) _UpperCamelCase : List[Any] = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[Any] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question=_snake_case , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : str = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : Dict = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [[{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''' ) def _lowercase ( self ) -> List[Any]: pass
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> bool: # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> bool: # Base Case if curr_ind == len(UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 ,len(UpperCamelCase ) ): if valid_connection(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ): # Insert current vertex into path as next transition _UpperCamelCase : Any = next_ver # Validate created path if util_hamilton_cycle(UpperCamelCase ,UpperCamelCase ,curr_ind + 1 ): return True # Backtrack _UpperCamelCase : Tuple = -1 return False def snake_case__ ( UpperCamelCase ,UpperCamelCase = 0 ) -> list[int]: _UpperCamelCase : List[str] = [-1] * (len(UpperCamelCase ) + 1) # initialize start and end of path with starting index _UpperCamelCase : str = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(UpperCamelCase ,UpperCamelCase ,1 ) else []
683
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers _UpperCAmelCase : Tuple = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) print("""NCCL version:""", torch.cuda.nccl.version()) except ImportError: print("""Torch version:""", None) try: import deepspeed print("""DeepSpeed version:""", deepspeed.__version__) except ImportError: print("""DeepSpeed version:""", None) try: import tensorflow as tf print("""TensorFlow version:""", tf.__version__) print("""TF GPUs available:""", bool(tf.config.list_physical_devices("""GPU"""))) print("""Number of TF GPUs available:""", len(tf.config.list_physical_devices("""GPU"""))) except ImportError: print("""TensorFlow version:""", None)
683
1
'''simple docstring''' from graphs.minimum_spanning_tree_kruskal import kruskal def snake_case__ ( ) -> Any: _UpperCamelCase : Optional[int] = 9 _UpperCamelCase : str = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] _UpperCamelCase : List[Any] = kruskal(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[int] = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(UpperCamelCase ) == sorted(UpperCamelCase )
683
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: return params[f'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase="attention" ) -> List[str]: _UpperCamelCase : Dict = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) _UpperCamelCase : int = k_tmp.reshape(k_tmp.shape[0] ,k_tmp.shape[1] * k_tmp.shape[2] ) _UpperCamelCase : str = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) _UpperCamelCase : Tuple = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] ,o_tmp.shape[2] ) _UpperCamelCase : Any = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) _UpperCamelCase : Optional[int] = q_tmp.reshape(q_tmp.shape[0] ,q_tmp.shape[1] * q_tmp.shape[2] ) _UpperCamelCase : Optional[Any] = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) _UpperCamelCase : List[Any] = v_tmp.reshape(v_tmp.shape[0] ,v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[str]: if split_mlp_wi: _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] _UpperCamelCase : Tuple = params[f'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] _UpperCamelCase : Optional[Any] = (wi_a, wi_a) else: _UpperCamelCase : str = params[f'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: return params[f'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def snake_case__ ( UpperCamelCase ,*, UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ) -> int: _UpperCamelCase : Any = traverse_util.flatten_dict(variables['''target'''] ) _UpperCamelCase : Optional[Any] = {'''/'''.join(UpperCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _UpperCamelCase : str = '''encoder/encoder/mlp/wi_0/kernel''' in old print('''Split MLP:''' ,UpperCamelCase ) _UpperCamelCase : Optional[int] = collections.OrderedDict() # Shared embeddings. _UpperCamelCase : str = old['''token_embedder/embedding'''] # Encoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''attention''' ) _UpperCamelCase : Tuple = layer_norm _UpperCamelCase : int = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : Dict = v.T # Block i, layer 1 (MLP). _UpperCamelCase : Dict = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : int = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,UpperCamelCase ) _UpperCamelCase : Union[str, Any] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Optional[Any] = wi[1].T else: _UpperCamelCase : List[Any] = wi.T _UpperCamelCase : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : Union[str, Any] = tax_relpos_bias_lookup( UpperCamelCase ,UpperCamelCase ,'''encoder''' ).T _UpperCamelCase : List[str] = old['''encoder/encoder_norm/scale'''] if not scalable_attention: _UpperCamelCase : List[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''encoder''' ).T _UpperCamelCase : Optional[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''decoder''' ).T if not is_encoder_only: # Decoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_self_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''self_attention''' ) _UpperCamelCase : int = layer_norm _UpperCamelCase : Union[str, Any] = k.T _UpperCamelCase : Optional[int] = o.T _UpperCamelCase : Dict = q.T _UpperCamelCase : Tuple = v.T # Block i, layer 1 (Cross Attention). _UpperCamelCase : str = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_cross_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Dict = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''encoder_decoder_attention''' ) _UpperCamelCase : Dict = layer_norm _UpperCamelCase : Optional[int] = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : str = v.T # Block i, layer 2 (MLP). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : List[Any] = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,UpperCamelCase ) _UpperCamelCase : List[str] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Union[str, Any] = wi[1].T else: _UpperCamelCase : Dict = wi.T _UpperCamelCase : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : int = tax_relpos_bias_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ).T _UpperCamelCase : Optional[int] = old['''decoder/decoder_norm/scale'''] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _UpperCamelCase : str = old['''decoder/logits_dense/kernel'''].T return new def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Optional[int]: _UpperCamelCase : str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : str = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : int = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''' ) _UpperCamelCase : Any = state_dict['''shared.weight'''] return state_dict def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Any: _UpperCamelCase : List[Any] = checkpoints.load_tax_checkpoint(UpperCamelCase ) _UpperCamelCase : str = convert_tax_to_pytorch( UpperCamelCase ,num_layers=config.num_layers ,is_encoder_only=UpperCamelCase ,scalable_attention=UpperCamelCase ) _UpperCamelCase : Optional[Any] = make_state_dict(UpperCamelCase ,UpperCamelCase ) model.load_state_dict(UpperCamelCase ,strict=UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ,UpperCamelCase = False ,) -> int: _UpperCamelCase : int = MTaConfig.from_json_file(UpperCamelCase ) print(f'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _UpperCamelCase : Optional[int] = UMTaEncoderModel(UpperCamelCase ) else: _UpperCamelCase : Optional[int] = UMTaForConditionalGeneration(UpperCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(UpperCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(UpperCamelCase ) print('''Done''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
683
1
'''simple docstring''' from __future__ import annotations def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,) -> tuple: if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' from __future__ import annotations from functools import lru_cache from math import ceil _UpperCAmelCase : int = 100 _UpperCAmelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _UpperCAmelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_00 ) def snake_case__ ( UpperCamelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _UpperCamelCase : set[int] = set() _UpperCamelCase : int _UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def snake_case__ ( UpperCamelCase = 50_00 ) -> int | None: for number_to_partition in range(1 ,UpperCamelCase ): if len(partition(UpperCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
683
1
'''simple docstring''' import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : int = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } _UpperCAmelCase : Tuple = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: for attribute in key.split('''.''' ): _UpperCamelCase : Union[str, Any] = getattr(UpperCamelCase ,UpperCamelCase ) if weight_type is not None: _UpperCamelCase : Optional[int] = getattr(UpperCamelCase ,UpperCamelCase ).shape else: _UpperCamelCase : List[str] = hf_pointer.shape assert hf_shape == value.shape, ( f'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": _UpperCamelCase : Any = value elif weight_type == "weight_g": _UpperCamelCase : List[str] = value elif weight_type == "weight_v": _UpperCamelCase : Optional[int] = value elif weight_type == "bias": _UpperCamelCase : Optional[Any] = value else: _UpperCamelCase : int = value logger.info(f'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : str = [] _UpperCamelCase : Union[str, Any] = fairseq_model.state_dict() _UpperCamelCase : Optional[int] = hf_model.feature_extractor _UpperCamelCase : int = hf_model.adapter for name, value in fairseq_dict.items(): _UpperCamelCase : Optional[int] = False if "conv_layers" in name: load_conv_layer( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,hf_model.config.feat_extract_norm == '''group''' ,) _UpperCamelCase : str = True elif any(x in name for x in ['''adaptor''', '''w2v_encoder.proj.''', '''w2v_proj_ln.'''] ): load_adapter(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : str = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _UpperCamelCase : List[Any] = True if "*" in mapped_key: _UpperCamelCase : int = name.split(UpperCamelCase )[0].split('''.''' )[-2] _UpperCamelCase : Optional[int] = mapped_key.replace('''*''' ,UpperCamelCase ) if "weight_g" in name: _UpperCamelCase : Union[str, Any] = '''weight_g''' elif "weight_v" in name: _UpperCamelCase : Tuple = '''weight_v''' elif "bias" in name: _UpperCamelCase : List[str] = '''bias''' elif "weight" in name: _UpperCamelCase : int = '''weight''' else: _UpperCamelCase : Optional[Any] = None set_recursively(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) continue if not is_used: unused_weights.append(UpperCamelCase ) logger.warning(f'''Unused weights: {unused_weights}''' ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: _UpperCamelCase : Optional[int] = full_name.split('''conv_layers.''' )[-1] _UpperCamelCase : Any = name.split('''.''' ) _UpperCamelCase : List[str] = int(items[0] ) _UpperCamelCase : List[str] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) _UpperCamelCase : int = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) _UpperCamelCase : Tuple = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) _UpperCamelCase : Optional[Any] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) _UpperCamelCase : Optional[Any] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Optional[Any]: _UpperCamelCase : int = full_name.split('''adaptor.''' )[-1] _UpperCamelCase : Dict = name.split('''.''' ) if items[1].isdigit(): _UpperCamelCase : str = int(items[1] ) else: _UpperCamelCase : List[Any] = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.''' _UpperCamelCase : Any = value logger.info(f'''Adapter proj layer norm bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.''' _UpperCamelCase : Any = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.''' _UpperCamelCase : Dict = value logger.info(f'''Adapter proj layer bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.''' _UpperCamelCase : Optional[int] = value logger.info(f'''Adapter proj layer weight was initialized from {full_name}.''' ) elif isinstance(UpperCamelCase ,UpperCamelCase ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.''' _UpperCamelCase : Any = value logger.info(f'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), f'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.''' _UpperCamelCase : List[str] = value logger.info(f'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) else: unused_weights.append(UpperCamelCase ) def snake_case__ ( UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase, _UpperCamelCase : List[Any] = emb.weight.shape _UpperCamelCase : List[str] = nn.Linear(UpperCamelCase ,UpperCamelCase ,bias=UpperCamelCase ) _UpperCamelCase : Optional[int] = emb.weight.data return lin_layer @torch.no_grad() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,) -> Optional[Any]: _UpperCamelCase : Dict = WavaVecaConfig.from_pretrained( UpperCamelCase ,add_adapter=UpperCamelCase ,adapter_stride=UpperCamelCase ,adapter_kernel_size=UpperCamelCase ,use_auth_token=UpperCamelCase ,output_hidden_size=UpperCamelCase ,) _UpperCamelCase : str = MBartConfig.from_pretrained(UpperCamelCase ) # load model _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={ '''config_yaml''': config_yaml_path, '''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path, '''load_pretrained_decoder_from''': None, } ,) _UpperCamelCase : Union[str, Any] = model[0].eval() # load feature extractor _UpperCamelCase : List[Any] = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase ,use_auth_token=UpperCamelCase ) # set weights for wav2vec2 encoder _UpperCamelCase : Union[str, Any] = WavaVecaModel(UpperCamelCase ) recursively_load_weights_wavaveca(model.encoder ,UpperCamelCase ) # load decoder weights _UpperCamelCase : int = MBartForCausalLM(UpperCamelCase ) _UpperCamelCase, _UpperCamelCase : Optional[int] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() ,strict=UpperCamelCase ) logger.warning(f'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(f'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) _UpperCamelCase : Optional[Any] = SpeechEncoderDecoderModel(encoder=UpperCamelCase ,decoder=UpperCamelCase ) _UpperCamelCase : str = False _UpperCamelCase : List[Any] = MBartaaTokenizer(UpperCamelCase ) tokenizer.save_pretrained(UpperCamelCase ) _UpperCamelCase : Dict = hf_wavavec.config.to_dict() _UpperCamelCase : Tuple = tokenizer.pad_token_id _UpperCamelCase : int = tokenizer.bos_token_id _UpperCamelCase : Dict = tokenizer.eos_token_id _UpperCamelCase : Dict = '''mbart50''' _UpperCamelCase : str = '''wav2vec2''' _UpperCamelCase : Optional[Any] = tokenizer.eos_token_id _UpperCamelCase : Union[str, Any] = 25_00_04 _UpperCamelCase : List[Any] = tokenizer.eos_token_id _UpperCamelCase : Optional[int] = SpeechEncoderDecoderConfig.from_dict(UpperCamelCase ) hf_wavavec.save_pretrained(UpperCamelCase ) feature_extractor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_yaml_path""", default=None, type=str, help="""Path to yaml file of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-xls-r-1b""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/mbart-large-50-one-to-many-mmt""", type=str, help="""Path to hf decoder checkpoint config""", ) parser.add_argument("""--add_adapter""", default=True, type=bool, help="""whethere to add model adapter layers""") parser.add_argument("""--adapter_stride""", default=2, type=int, help="""stride of adapter layers""") parser.add_argument("""--adapter_kernel_size""", default=3, type=int, help="""kernel size of adapter layers""") parser.add_argument("""--encoder_output_dim""", default=1024, type=int, help="""encoder output dim""") parser.add_argument("""--start_token_id""", default=250004, type=int, help="""`decoder_start_token_id` of model config""") _UpperCAmelCase : Optional[int] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
683
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _UpperCAmelCase : Dict = """bart""" _UpperCAmelCase : List[str] = True @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> int: if LOAD_DENSE_INDEX: _UpperCamelCase : Optional[int] = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _UpperCamelCase : Optional[Any] = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _UpperCamelCase : Tuple = qar_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = (None, None) if MODEL_TYPE == "bart": _UpperCamelCase : Any = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _UpperCamelCase : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _UpperCamelCase : Dict = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _UpperCamelCase : Tuple = sas_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = make_qa_sas_model( model_name='''t5-small''' ,from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' ,device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> List[Any]: if LOAD_DENSE_INDEX: _UpperCamelCase : str = faiss.StandardGpuResources() _UpperCamelCase : Optional[int] = datasets.load_dataset(path='''wiki_snippets''' ,name='''wiki40b_en_100_0''' )['''train'''] _UpperCamelCase : List[str] = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(wikiaab_passages.num_rows, 1_28) ,) _UpperCamelCase : Any = faiss.IndexFlatIP(1_28 ) _UpperCamelCase : str = faiss.index_cpu_to_gpu(UpperCamelCase ,1 ,UpperCamelCase ) wikiaab_gpu_index_flat.add(UpperCamelCase ) # TODO fix for larger GPU else: _UpperCamelCase, _UpperCamelCase : Optional[int] = (None, None) _UpperCamelCase : int = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : List[Any] = datasets.load_dataset('''eli5''' ,name='''LFQA_reddit''' ) _UpperCamelCase : Optional[int] = elia['''train_eli5'''] _UpperCamelCase : Any = np.memmap( '''eli5_questions_reps.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(elia_train.num_rows, 1_28) ) _UpperCamelCase : Optional[Any] = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(UpperCamelCase ) return (elia_train, eli5_train_q_index) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_indexes() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_models() _UpperCAmelCase , _UpperCAmelCase : int = load_train_data() def snake_case__ ( UpperCamelCase ,UpperCamelCase=10 ) -> Optional[Any]: _UpperCamelCase : Optional[int] = embed_questions_for_retrieval([question] ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase, _UpperCamelCase : Optional[Any] = eli5_train_q_index.search(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = [elia_train[int(UpperCamelCase )] for i in I[0]] return nn_examples def snake_case__ ( UpperCamelCase ,UpperCamelCase="wiki40b" ,UpperCamelCase="dense" ,UpperCamelCase=10 ) -> Optional[int]: if source == "none": _UpperCamelCase, _UpperCamelCase : Dict = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _UpperCamelCase, _UpperCamelCase : str = query_qa_dense_index( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else: _UpperCamelCase, _UpperCamelCase : str = query_es_index( UpperCamelCase ,UpperCamelCase ,index_name='''english_wiki40b_snippets_100w''' ,n_results=UpperCamelCase ,) _UpperCamelCase : Optional[int] = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _UpperCamelCase : Optional[Any] = '''question: {} context: {}'''.format(UpperCamelCase ,UpperCamelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda UpperCamelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda UpperCamelCase : None), } ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=64 ,UpperCamelCase=2_56 ,UpperCamelCase=False ,UpperCamelCase=2 ,UpperCamelCase=0.95 ,UpperCamelCase=0.8 ) -> Optional[Any]: with torch.no_grad(): _UpperCamelCase : Any = qa_sas_generate( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,num_answers=1 ,num_beams=UpperCamelCase ,min_len=UpperCamelCase ,max_len=UpperCamelCase ,do_sample=UpperCamelCase ,temp=UpperCamelCase ,top_p=UpperCamelCase ,top_k=UpperCamelCase ,max_input_length=10_24 ,device='''cuda:0''' ,)[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar _UpperCAmelCase : str = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _UpperCAmelCase : Tuple = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _UpperCAmelCase : Dict = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _UpperCAmelCase : List[str] = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _UpperCAmelCase : Optional[int] = st.sidebar.checkbox("""Demo options""") if demo_options: _UpperCAmelCase : List[str] = st.sidebar.selectbox( """""", action_list, index=3, ) _UpperCAmelCase : List[Any] = action_list.index(action_st) _UpperCAmelCase : Tuple = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) _UpperCAmelCase : Optional[Any] = show_type == """Show full text of passages""" else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : str = True _UpperCAmelCase : str = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: _UpperCAmelCase : Optional[Any] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) _UpperCAmelCase : Optional[Any] = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: _UpperCAmelCase : Dict = """wiki40b""" _UpperCAmelCase : str = """dense""" _UpperCAmelCase : List[str] = """beam""" _UpperCAmelCase : Dict = 2 _UpperCAmelCase : List[str] = 64 _UpperCAmelCase : List[Any] = 256 _UpperCAmelCase : Tuple = None _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : int = st.sidebar.checkbox("""Generation options""") if generate_options: _UpperCAmelCase : Union[str, Any] = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) _UpperCAmelCase : Dict = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) _UpperCAmelCase : List[Any] = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": _UpperCAmelCase : List[str] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[int] = None # start main text _UpperCAmelCase : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _UpperCAmelCase : int = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": _UpperCAmelCase : Any = st.text_input("""Enter your question here:""", """""") else: _UpperCAmelCase : Tuple = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": _UpperCAmelCase , _UpperCAmelCase : str = make_support(question, source=wiki_source, method="""dense""", n_results=10) _UpperCAmelCase , _UpperCAmelCase : List[Any] = make_support(question, source=wiki_source, method="""sparse""", n_results=10) _UpperCAmelCase : int = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _UpperCAmelCase : int = support_list[:10] _UpperCAmelCase : Tuple = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _UpperCAmelCase , _UpperCAmelCase : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): _UpperCAmelCase : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) _UpperCAmelCase : List[Any] = res[1].strip() if sec_titles == "": _UpperCAmelCase : Optional[int] = """[{}]({})""".format(res[0], wiki_url) else: _UpperCAmelCase : Optional[int] = sec_titles.split(""" & """) _UpperCAmelCase : Tuple = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: _UpperCAmelCase : Dict = find_nearest_training(question) _UpperCAmelCase : List[Any] = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) _UpperCAmelCase : List[Any] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) _UpperCAmelCase : List[Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
683
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Tuple = DanceDiffusionPipeline A__ : Union[str, Any] = UNCONDITIONAL_AUDIO_GENERATION_PARAMS A__ : Union[str, Any] = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } A__ : Optional[int] = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS A__ : List[Any] = False A__ : Dict = False def _lowercase ( self ) -> int: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDModel( block_out_channels=(32, 32, 64) , extra_in_channels=16 , sample_size=512 , sample_rate=16000 , in_channels=2 , out_channels=2 , flip_sin_to_cos=_snake_case , use_timestep_embedding=_snake_case , time_embedding_type='''fourier''' , mid_block_type='''UNetMidBlock1D''' , down_block_types=('''DownBlock1DNoSkip''', '''DownBlock1D''', '''AttnDownBlock1D''') , up_block_types=('''AttnUpBlock1D''', '''UpBlock1D''', '''UpBlock1DNoSkip''') , ) _UpperCamelCase : List[Any] = IPNDMScheduler() _UpperCamelCase : List[str] = { '''unet''': unet, '''scheduler''': scheduler, } return components def _lowercase ( self , _snake_case , _snake_case=0 ) -> List[str]: if str(_snake_case ).startswith('''mps''' ): _UpperCamelCase : Union[str, Any] = torch.manual_seed(_snake_case ) else: _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) _UpperCamelCase : Union[str, Any] = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 4, } return inputs def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Optional[Any] = self.get_dummy_components() _UpperCamelCase : Optional[int] = DanceDiffusionPipeline(**_snake_case ) _UpperCamelCase : Any = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Union[str, Any] = self.get_dummy_inputs(_snake_case ) _UpperCamelCase : Dict = pipe(**_snake_case ) _UpperCamelCase : int = output.audios _UpperCamelCase : Optional[Any] = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) _UpperCamelCase : str = np.array([-0.7_265, 1.0_000, -0.8_388, 0.1_175, 0.9_498, -1.0_000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def _lowercase ( self ) -> Any: return super().test_save_load_local() @skip_mps def _lowercase ( self ) -> Dict: return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) @skip_mps def _lowercase ( self ) -> Any: return super().test_save_load_optional_components() @skip_mps def _lowercase ( self ) -> List[str]: return super().test_attention_slicing_forward_pass() def _lowercase ( self ) -> Any: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> int: _UpperCamelCase : Tuple = torch_device _UpperCamelCase : str = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' ) _UpperCamelCase : Any = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : str = torch.manual_seed(0 ) _UpperCamelCase : Dict = pipe(generator=_snake_case , num_inference_steps=100 , audio_length_in_s=4.096 ) _UpperCamelCase : List[str] = output.audios _UpperCamelCase : Union[str, Any] = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _UpperCamelCase : Any = np.array([-0.0_192, -0.0_231, -0.0_318, -0.0_059, 0.0_002, -0.0_020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self ) -> int: _UpperCamelCase : List[str] = torch_device _UpperCamelCase : int = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' , torch_dtype=torch.floataa ) _UpperCamelCase : List[str] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Dict = torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipe(generator=_snake_case , num_inference_steps=100 , audio_length_in_s=4.096 ) _UpperCamelCase : Optional[Any] = output.audios _UpperCamelCase : Tuple = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _UpperCamelCase : int = np.array([-0.0_367, -0.0_488, -0.0_771, -0.0_525, -0.0_444, -0.0_341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2
683
'''simple docstring''' from collections.abc import Iterable from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> Optional[int]: _UpperCamelCase : int = value _UpperCamelCase : Node | None = None # Added in order to delete a node easier _UpperCamelCase : Node | None = None _UpperCamelCase : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> List[Any]: _UpperCamelCase : str = root def __str__( self ) -> str: return str(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if new_children is not None: # reset its kids _UpperCamelCase : Union[str, Any] = node.parent if node.parent is not None: # reset its parent if self.is_right(_snake_case ): # If it is the right children _UpperCamelCase : str = new_children else: _UpperCamelCase : Any = new_children else: _UpperCamelCase : Any = new_children def _lowercase ( self , _snake_case ) -> bool: if node.parent and node.parent.right: return node == node.parent.right return False def _lowercase ( self ) -> bool: return self.root is None def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : List[Any] = Node(_snake_case ) # create a new Node if self.empty(): # if Tree is empty _UpperCamelCase : Optional[Any] = new_node # set its root else: # Tree is not empty _UpperCamelCase : int = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: _UpperCamelCase : Union[str, Any] = new_node # We insert the new node in a leaf break else: _UpperCamelCase : Union[str, Any] = parent_node.left else: if parent_node.right is None: _UpperCamelCase : Any = new_node break else: _UpperCamelCase : str = parent_node.right _UpperCamelCase : Any = parent_node def _lowercase ( self , *_snake_case ) -> None: for value in values: self.__insert(_snake_case ) def _lowercase ( self , _snake_case ) -> Node | None: if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: _UpperCamelCase : List[str] = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: _UpperCamelCase : Optional[Any] = node.left if value < node.value else node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: if self.root is None: return None _UpperCamelCase : Dict = self.root if not self.empty(): while node.right is not None: _UpperCamelCase : Tuple = node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: _UpperCamelCase : Optional[Any] = self.root if self.root is None: return None if not self.empty(): _UpperCamelCase : Optional[int] = self.root while node.left is not None: _UpperCamelCase : List[str] = node.left return node def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : str = self.search(_snake_case ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_snake_case , _snake_case ) elif node.left is None: # Has only right children self.__reassign_nodes(_snake_case , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_snake_case , node.left ) else: _UpperCamelCase : List[str] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore _UpperCamelCase : int = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _lowercase ( self , _snake_case ) -> Iterable: if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _lowercase ( self , _snake_case=None ) -> Any: if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if node: self.inorder(_snake_case , node.left ) arr.append(node.value ) self.inorder(_snake_case , node.right ) def _lowercase ( self , _snake_case , _snake_case ) -> int: _UpperCamelCase : list[int] = [] self.inorder(_snake_case , _snake_case ) # append all values to list using inorder traversal return arr[k - 1] def snake_case__ ( UpperCamelCase ) -> list[Node]: _UpperCamelCase : int = [] if curr_node is not None: _UpperCamelCase : Any = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def snake_case__ ( ) -> None: _UpperCamelCase : Any = (8, 3, 6, 1, 10, 14, 13, 4, 7) _UpperCamelCase : Tuple = BinarySearchTree() for i in testlist: t.insert(UpperCamelCase ) # Prints all the elements of the list in order traversal print(UpperCamelCase ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' ,t.get_max().value ) # type: ignore print('''Min Value: ''' ,t.get_min().value ) # type: ignore for i in testlist: t.remove(UpperCamelCase ) print(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
683
1
'''simple docstring''' import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : str = OpenAIGPTTokenizer A__ : int = OpenAIGPTTokenizerFast A__ : str = True A__ : Optional[Any] = False def _lowercase ( self ) -> Union[str, Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCamelCase : int = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] _UpperCamelCase : str = dict(zip(_snake_case , range(len(_snake_case ) ) ) ) _UpperCamelCase : int = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] _UpperCamelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCamelCase : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(_snake_case ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(_snake_case ) ) def _lowercase ( self , _snake_case ) -> Tuple: return "lower newer", "lower newer" def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) _UpperCamelCase : Union[str, Any] = '''lower''' _UpperCamelCase : List[str] = ['''low''', '''er</w>'''] _UpperCamelCase : Dict = tokenizer.tokenize(_snake_case ) self.assertListEqual(_snake_case , _snake_case ) _UpperCamelCase : str = tokens + ['''<unk>'''] _UpperCamelCase : Any = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(_snake_case ) , _snake_case ) def _lowercase ( self , _snake_case=15 ) -> Dict: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): _UpperCamelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_snake_case , **_snake_case ) # Simple input _UpperCamelCase : Tuple = '''This is a simple input''' _UpperCamelCase : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] _UpperCamelCase : Dict = ('''This is a simple input''', '''This is a pair''') _UpperCamelCase : Tuple = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(_snake_case , tokenizer_r.encode , _snake_case , max_length=_snake_case , padding='''max_length''' ) # Simple input self.assertRaises(_snake_case , tokenizer_r.encode_plus , _snake_case , max_length=_snake_case , padding='''max_length''' ) # Simple input self.assertRaises( _snake_case , tokenizer_r.batch_encode_plus , _snake_case , max_length=_snake_case , padding='''max_length''' , ) # Pair input self.assertRaises(_snake_case , tokenizer_r.encode , _snake_case , max_length=_snake_case , padding='''max_length''' ) # Pair input self.assertRaises(_snake_case , tokenizer_r.encode_plus , _snake_case , max_length=_snake_case , padding='''max_length''' ) # Pair input self.assertRaises( _snake_case , tokenizer_r.batch_encode_plus , _snake_case , max_length=_snake_case , padding='''max_length''' , ) def _lowercase ( self ) -> Any: pass @require_ftfy @require_spacy @require_tokenizers class UpperCAmelCase ( a_ ): """simple docstring""" pass
683
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off _UpperCAmelCase : Dict = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 10563, 10786, 11420, 11709, 11907, 13163, 13697, 13700, 14808, 15306, 16410, 16791, 17992, 19203, 19510, 20724, 22305, 22935, 27007, 30109, 30420, 33409, 34949, 40283, 40493, 40549, 47282, 49146, 50257, 50359, 50360, 50361 ] _UpperCAmelCase : int = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362 ] class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = 'whisper' A__ : Tuple = ['past_key_values'] A__ : Optional[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , _snake_case=51865 , _snake_case=80 , _snake_case=6 , _snake_case=4 , _snake_case=6 , _snake_case=4 , _snake_case=1536 , _snake_case=1536 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=50257 , _snake_case=True , _snake_case=True , _snake_case="gelu" , _snake_case=256 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.02 , _snake_case=False , _snake_case=1500 , _snake_case=448 , _snake_case=50256 , _snake_case=50256 , _snake_case=50256 , _snake_case=None , _snake_case=[220, 50256] , _snake_case=False , _snake_case=256 , _snake_case=False , _snake_case=0.05 , _snake_case=10 , _snake_case=2 , _snake_case=0.0 , _snake_case=10 , _snake_case=0 , _snake_case=7 , **_snake_case , ) -> Any: _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Union[str, Any] = num_mel_bins _UpperCamelCase : List[str] = d_model _UpperCamelCase : str = encoder_layers _UpperCamelCase : Optional[int] = encoder_attention_heads _UpperCamelCase : str = decoder_layers _UpperCamelCase : Tuple = decoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : Optional[int] = encoder_ffn_dim _UpperCamelCase : Any = dropout _UpperCamelCase : Optional[Any] = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : int = activation_function _UpperCamelCase : List[Any] = init_std _UpperCamelCase : Optional[int] = encoder_layerdrop _UpperCamelCase : str = decoder_layerdrop _UpperCamelCase : List[str] = use_cache _UpperCamelCase : Optional[Any] = encoder_layers _UpperCamelCase : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : List[str] = max_source_positions _UpperCamelCase : Optional[Any] = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. _UpperCamelCase : str = classifier_proj_size _UpperCamelCase : List[str] = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCamelCase : int = apply_spec_augment _UpperCamelCase : str = mask_time_prob _UpperCamelCase : int = mask_time_length _UpperCamelCase : List[Any] = mask_time_min_masks _UpperCamelCase : List[str] = mask_feature_prob _UpperCamelCase : Optional[int] = mask_feature_length _UpperCamelCase : Union[str, Any] = mask_feature_min_masks _UpperCamelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , is_encoder_decoder=_snake_case , decoder_start_token_id=_snake_case , suppress_tokens=_snake_case , begin_suppress_tokens=_snake_case , **_snake_case , ) class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: _UpperCamelCase : Dict = OrderedDict( [ ('''input_features''', {0: '''batch''', 1: '''feature_size''', 2: '''encoder_sequence'''}), ] ) if self.use_past: _UpperCamelCase : Tuple = {0: '''batch'''} else: _UpperCamelCase : Dict = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_snake_case , direction='''inputs''' ) return common_inputs def _lowercase ( self , _snake_case , _snake_case = -1 , _snake_case = -1 , _snake_case = False , _snake_case = None , _snake_case = 22050 , _snake_case = 5.0 , _snake_case = 220 , ) -> Mapping[str, Any]: _UpperCamelCase : Optional[int] = OrderedDict() _UpperCamelCase : Tuple = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_snake_case , framework=_snake_case , sampling_rate=_snake_case , time_duration=_snake_case , frequency=_snake_case , ) _UpperCamelCase : int = encoder_inputs['''input_features'''].shape[2] _UpperCamelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length _UpperCamelCase : str = super().generate_dummy_inputs( preprocessor.tokenizer , _snake_case , _snake_case , _snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = encoder_inputs.pop('''input_features''' ) _UpperCamelCase : Dict = decoder_inputs.pop('''decoder_input_ids''' ) if "past_key_values" in decoder_inputs: _UpperCamelCase : List[str] = decoder_inputs.pop('''past_key_values''' ) return dummy_inputs @property def _lowercase ( self ) -> float: return 1E-3
683
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Tuple = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : str = { """vocab_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """yjernite/retribert-base-uncased""": 512, } _UpperCAmelCase : Optional[Any] = { """yjernite/retribert-base-uncased""": {"""do_lower_case""": True}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Union[str, Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Dict = RetriBertTokenizer A__ : List[Any] = ['input_ids', 'attention_mask'] def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> Optional[int]: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : Optional[int] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : Tuple = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Union[str, Any] = do_lower_case _UpperCamelCase : Optional[int] = strip_accents _UpperCamelCase : Optional[int] = tokenize_chinese_chars _UpperCamelCase : str = normalizer_class(**_snake_case ) _UpperCamelCase : Any = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Dict: _UpperCamelCase : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Dict = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Union[str, Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
'''simple docstring''' import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser( description=( """Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""roberta""", choices=["""roberta""", """gpt2"""]) parser.add_argument("""--model_name""", default="""roberta-large""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_roberta_048131723.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") _UpperCAmelCase : int = parser.parse_args() if args.model_type == "roberta": _UpperCAmelCase : Union[str, Any] = RobertaForMaskedLM.from_pretrained(args.model_name) _UpperCAmelCase : int = """roberta""" elif args.model_type == "gpt2": _UpperCAmelCase : Optional[int] = GPTaLMHeadModel.from_pretrained(args.model_name) _UpperCAmelCase : Optional[int] = """transformer""" _UpperCAmelCase : Tuple = model.state_dict() _UpperCAmelCase : int = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: _UpperCAmelCase : Optional[Any] = state_dict[f"""{prefix}.{param_name}"""] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: _UpperCAmelCase : Tuple = f"""{prefix}.embeddings.{w}.weight""" _UpperCAmelCase : Optional[Any] = state_dict[param_name] for w in ["weight", "bias"]: _UpperCAmelCase : Union[str, Any] = f"""{prefix}.embeddings.LayerNorm.{w}""" _UpperCAmelCase : str = state_dict[param_name] # Transformer Blocks # _UpperCAmelCase : Dict = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: _UpperCAmelCase : str = state_dict[ f"""{prefix}.h.{teacher_idx}.{layer}.{w}""" ] _UpperCAmelCase : Any = state_dict[f"""{prefix}.h.{teacher_idx}.attn.bias"""] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: _UpperCAmelCase : Optional[Any] = state_dict[ f"""{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}""" ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: _UpperCAmelCase : Dict = state_dict[f"""{layer}"""] if args.vocab_transform: for w in ["weight", "bias"]: _UpperCAmelCase : int = state_dict[f"""lm_head.dense.{w}"""] _UpperCAmelCase : int = state_dict[f"""lm_head.layer_norm.{w}"""] elif args.model_type == "gpt2": for w in ["weight", "bias"]: _UpperCAmelCase : List[str] = state_dict[f"""{prefix}.ln_f.{w}"""] _UpperCAmelCase : Any = state_dict["""lm_head.weight"""] print(f"""N layers selected for distillation: {std_idx}""") print(f"""Number of params transferred for distillation: {len(compressed_sd.keys())}""") print(f"""Save transferred checkpoint to {args.dump_checkpoint}.""") torch.save(compressed_sd, args.dump_checkpoint)
683
1
'''simple docstring''' import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Dict = 'ssube/stable-diffusion-x4-upscaler-onnx' def _lowercase ( self , _snake_case=0 ) -> str: _UpperCamelCase : int = floats_tensor((1, 3, 128, 128) , rng=random.Random(_snake_case ) ) _UpperCamelCase : Optional[Any] = torch.manual_seed(_snake_case ) _UpperCamelCase : int = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : str = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : int = self.get_dummy_inputs() _UpperCamelCase : Dict = pipe(**_snake_case ).images _UpperCamelCase : int = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Dict = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def _lowercase ( self ) -> str: _UpperCamelCase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) _UpperCamelCase : Optional[int] = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Union[str, Any] = self.get_dummy_inputs() _UpperCamelCase : List[Any] = pipe(**_snake_case ).images _UpperCamelCase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : str = np.array( [0.6_898_892, 0.59_240_556, 0.52_499_527, 0.58_866_215, 0.52_258_235, 0.52_572_715, 0.62_414_473, 0.6_174_387, 0.6_214_964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) _UpperCamelCase : int = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Optional[Any] = self.get_dummy_inputs() _UpperCamelCase : List[Any] = pipe(**_snake_case ).images _UpperCamelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : List[Any] = np.array( [0.7_659_278, 0.76_437_664, 0.75_579_107, 0.7_691_116, 0.77_666_986, 0.7_727_672, 0.7_758_664, 0.7_812_226, 0.76_942_515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self ) -> List[str]: _UpperCamelCase : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) _UpperCamelCase : List[Any] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : List[Any] = self.get_dummy_inputs() _UpperCamelCase : Any = pipe(**_snake_case ).images _UpperCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : int = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) _UpperCamelCase : List[Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : List[str] = self.get_dummy_inputs() _UpperCamelCase : List[Any] = pipe(**_snake_case ).images _UpperCamelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase : Any = np.array( [0.77_424_496, 0.773_601, 0.7_645_288, 0.7_769_598, 0.7_772_739, 0.7_738_688, 0.78_187_233, 0.77_879_584, 0.767_043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def _lowercase ( self ) -> Any: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _lowercase ( self ) -> Dict: _UpperCamelCase : Union[str, Any] = ort.SessionOptions() _UpperCamelCase : Dict = False return options def _lowercase ( self ) -> str: _UpperCamelCase : Union[str, Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) _UpperCamelCase : Optional[int] = init_image.resize((128, 128) ) # using the PNDM scheduler by default _UpperCamelCase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : List[str] = '''A fantasy landscape, trending on artstation''' _UpperCamelCase : List[str] = torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipe( prompt=_snake_case , image=_snake_case , guidance_scale=7.5 , num_inference_steps=10 , generator=_snake_case , output_type='''np''' , ) _UpperCamelCase : Dict = output.images _UpperCamelCase : Dict = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCamelCase : Union[str, Any] = np.array([0.4_883, 0.4_947, 0.4_980, 0.4_975, 0.4_982, 0.4_980, 0.5_000, 0.5_006, 0.4_972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowercase ( self ) -> Tuple: _UpperCamelCase : Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) _UpperCamelCase : Dict = init_image.resize((128, 128) ) _UpperCamelCase : Tuple = LMSDiscreteScheduler.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , subfolder='''scheduler''' ) _UpperCamelCase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , scheduler=_snake_case , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : List[Any] = '''A fantasy landscape, trending on artstation''' _UpperCamelCase : List[str] = torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = pipe( prompt=_snake_case , image=_snake_case , guidance_scale=7.5 , num_inference_steps=20 , generator=_snake_case , output_type='''np''' , ) _UpperCamelCase : Tuple = output.images _UpperCamelCase : Optional[int] = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCamelCase : Dict = np.array( [0.50_173_753, 0.50_223_356, 0.502_039, 0.50_233_036, 0.5_023_725, 0.5_022_601, 0.5_018_758, 0.50_234_085, 0.50_241_566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
683
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = None _UpperCamelCase : int = 20 _UpperCamelCase : Any = self._get_uniform_logits(batch_size=2 , length=_snake_case ) # tweak scores to not be uniform anymore _UpperCamelCase : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch _UpperCamelCase : Dict = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax _UpperCamelCase : Any = jax.nn.softmax(_snake_case , axis=-1 ) _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : List[str] = FlaxTemperatureLogitsWarper(temperature=1.3 ) _UpperCamelCase : List[str] = jax.nn.softmax(temp_dist_warper_sharper(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) _UpperCamelCase : str = jax.nn.softmax(temp_dist_warper_smoother(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _lowercase ( self ) -> Any: _UpperCamelCase : List[Any] = None _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : Any = 2 # create ramp distribution _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() _UpperCamelCase : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case _UpperCamelCase : Optional[int] = 5 _UpperCamelCase : str = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) _UpperCamelCase : Union[str, Any] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, length) ).copy() _UpperCamelCase : Optional[Any] = top_k_warp_safety_check(_snake_case , _snake_case , cur_len=_snake_case ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = None _UpperCamelCase : Any = 10 _UpperCamelCase : List[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) _UpperCamelCase : Tuple = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) _UpperCamelCase : List[str] = FlaxTopPLogitsWarper(0.8 ) _UpperCamelCase : Dict = np.exp(top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 _UpperCamelCase : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # check edge cases with negative and extreme logits _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme _UpperCamelCase : Tuple = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept _UpperCamelCase : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) _UpperCamelCase : Dict = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = 20 _UpperCamelCase : Optional[int] = 4 _UpperCamelCase : int = 0 _UpperCamelCase : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) # check that min length is applied at length 5 _UpperCamelCase : Any = ids_tensor((batch_size, 20) , vocab_size=20 ) _UpperCamelCase : int = 5 _UpperCamelCase : List[Any] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 _UpperCamelCase : Optional[int] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = 15 _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = 20 _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) # check that all scores are -inf except the bos_token_id score _UpperCamelCase : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) _UpperCamelCase : Optional[int] = 1 _UpperCamelCase : str = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : str = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 _UpperCamelCase : List[str] = 3 _UpperCamelCase : Tuple = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 20 _UpperCamelCase : Tuple = 4 _UpperCamelCase : Any = 0 _UpperCamelCase : str = 5 _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) # check that all scores are -inf except the eos_token_id when max_length is reached _UpperCamelCase : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) _UpperCamelCase : Dict = 4 _UpperCamelCase : Dict = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : int = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Any = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 4 _UpperCamelCase : Optional[Any] = 10 _UpperCamelCase : Dict = 15 _UpperCamelCase : Union[str, Any] = 2 _UpperCamelCase : Optional[Any] = 1 _UpperCamelCase : List[Any] = 15 # dummy input_ids and scores _UpperCamelCase : Optional[int] = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Any = input_ids.copy() _UpperCamelCase : int = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Tuple = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Optional[int] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : int = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : List[str] = 10 # no processor list _UpperCamelCase : Dict = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) # with processor list _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : Optional[Any] = processor(_snake_case , _snake_case , cur_len=_snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = 4 _UpperCamelCase : int = 10 _UpperCamelCase : List[Any] = 15 _UpperCamelCase : Dict = 2 _UpperCamelCase : Tuple = 1 _UpperCamelCase : Optional[int] = 15 # dummy input_ids and scores _UpperCamelCase : Tuple = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Optional[Any] = input_ids.copy() _UpperCamelCase : List[str] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[int] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : int = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : Union[str, Any] = 10 # no processor list def run_no_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : List[Any] = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) return scores # with processor list def run_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : List[str] = processor(_snake_case , _snake_case , cur_len=_snake_case ) return scores _UpperCamelCase : Dict = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jitted_run_no_processor_list(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = jitted_run_processor_list(_snake_case , _snake_case , _snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
683
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer _UpperCAmelCase : Dict = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast _UpperCAmelCase : List[Any] = TaTokenizerFast _UpperCAmelCase : str = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys _UpperCAmelCase : Any = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
683
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Optional[int] = pytest.mark.integration @pytest.mark.parametrize('''path''' ,['''paws''', '''csv'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: inspect_dataset(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' ,['''accuracy'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: inspect_metric(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[str] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[str]: with pytest.raises(UpperCamelCase ): get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) @pytest.mark.parametrize( '''path, expected''' ,[ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : int = get_dataset_config_names(UpperCamelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' ,[ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Dict = get_dataset_infos(UpperCamelCase ) assert list(infos.keys() ) == expected_configs _UpperCamelCase : Dict = expected_configs[0] assert expected_config in infos _UpperCamelCase : Any = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : List[Any] = get_dataset_infos(UpperCamelCase ) assert expected_config in infos _UpperCamelCase : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: with pytest.raises(UpperCamelCase ): get_dataset_split_names(UpperCamelCase ,config_name=UpperCamelCase )
683
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : str = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _UpperCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ) -> Dict: torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return model @property def _lowercase ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , cross_attention_dim=10 , ) return model @property def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') , up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') , ) _UpperCamelCase : int = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return vqvae, unet @slow def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Tuple = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _UpperCamelCase : int = DDPMScheduler() _UpperCamelCase : Optional[int] = AudioDiffusionPipeline(vqvae=_snake_case , unet=self.dummy_unet , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case , steps=4 ) _UpperCamelCase : Union[str, Any] = output.audios[0] _UpperCamelCase : Union[str, Any] = output.images[0] _UpperCamelCase : str = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : int = pipe(generator=_snake_case , steps=4 , return_dict=_snake_case ) _UpperCamelCase : int = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _UpperCamelCase : List[str] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : List[str] = np.frombuffer(image_from_tuple.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : int = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : str = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _UpperCamelCase : Dict = DDIMScheduler() _UpperCamelCase : str = self.dummy_vqvae_and_unet _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : Optional[Any] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _UpperCamelCase : Optional[Any] = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Tuple = pipe(raw_audio=_snake_case , generator=_snake_case , start_step=5 , steps=10 ) _UpperCamelCase : List[str] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _UpperCamelCase : Any = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Tuple = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : Any = self.dummy_unet_condition _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_snake_case , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : Union[str, Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : int = torch.rand((1, 1, 10) ) _UpperCamelCase : Optional[Any] = pipe(generator=_snake_case , encoding=_snake_case ) _UpperCamelCase : Dict = output.images[0] _UpperCamelCase : Dict = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Any = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = torch_device _UpperCamelCase : int = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) _UpperCamelCase : str = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case ) _UpperCamelCase : List[Any] = output.audios[0] _UpperCamelCase : List[Any] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _UpperCamelCase : Union[str, Any] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Union[str, Any] = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
683
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : List[str] = { """microsoft/focalnet-tiny""": """https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json""", } class UpperCAmelCase ( a_ , a_ ): """simple docstring""" A__ : Dict = 'focalnet' def __init__( self , _snake_case=224 , _snake_case=4 , _snake_case=3 , _snake_case=96 , _snake_case=False , _snake_case=[192, 384, 768, 768] , _snake_case=[2, 2, 6, 2] , _snake_case=[2, 2, 2, 2] , _snake_case=[3, 3, 3, 3] , _snake_case="gelu" , _snake_case=4.0 , _snake_case=0.0 , _snake_case=0.1 , _snake_case=False , _snake_case=1E-4 , _snake_case=False , _snake_case=False , _snake_case=False , _snake_case=0.02 , _snake_case=1E-5 , _snake_case=32 , _snake_case=None , _snake_case=None , **_snake_case , ) -> Union[str, Any]: super().__init__(**_snake_case ) _UpperCamelCase : Dict = image_size _UpperCamelCase : Optional[int] = patch_size _UpperCamelCase : Optional[int] = num_channels _UpperCamelCase : Tuple = embed_dim _UpperCamelCase : int = use_conv_embed _UpperCamelCase : List[str] = hidden_sizes _UpperCamelCase : Optional[int] = depths _UpperCamelCase : List[str] = focal_levels _UpperCamelCase : Dict = focal_windows _UpperCamelCase : Union[str, Any] = hidden_act _UpperCamelCase : List[Any] = mlp_ratio _UpperCamelCase : str = hidden_dropout_prob _UpperCamelCase : Optional[int] = drop_path_rate _UpperCamelCase : Optional[int] = use_layerscale _UpperCamelCase : Optional[int] = layerscale_value _UpperCamelCase : int = use_post_layernorm _UpperCamelCase : int = use_post_layernorm_in_modulation _UpperCamelCase : List[Any] = normalize_modulator _UpperCamelCase : int = initializer_range _UpperCamelCase : int = layer_norm_eps _UpperCamelCase : Optional[int] = encoder_stride _UpperCamelCase : Any = ['''stem'''] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] _UpperCamelCase, _UpperCamelCase : Optional[int] = get_aligned_output_features_output_indices( out_features=_snake_case , out_indices=_snake_case , stage_names=self.stage_names )
683
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_longt5""": ["""LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongT5Config""", """LongT5OnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongT5EncoderModel""", """LongT5ForConditionalGeneration""", """LongT5Model""", """LongT5PreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """FlaxLongT5ForConditionalGeneration""", """FlaxLongT5Model""", """FlaxLongT5PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
1
'''simple docstring''' from copy import deepcopy class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None , _snake_case = None ) -> None: if arr is None and size is not None: _UpperCamelCase : Optional[int] = size _UpperCamelCase : Dict = [0] * size elif arr is not None: self.init(_snake_case ) else: raise ValueError('''Either arr or size must be specified''' ) def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : int = len(_snake_case ) _UpperCamelCase : List[str] = deepcopy(_snake_case ) for i in range(1 , self.size ): _UpperCamelCase : str = self.next_(_snake_case ) if j < self.size: self.tree[j] += self.tree[i] def _lowercase ( self ) -> list[int]: _UpperCamelCase : Tuple = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): _UpperCamelCase : Optional[Any] = self.next_(_snake_case ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def _lowercase ( _snake_case ) -> int: return index + (index & (-index)) @staticmethod def _lowercase ( _snake_case ) -> int: return index - (index & (-index)) def _lowercase ( self , _snake_case , _snake_case ) -> None: if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value _UpperCamelCase : Union[str, Any] = self.next_(_snake_case ) def _lowercase ( self , _snake_case , _snake_case ) -> None: self.add(_snake_case , value - self.get(_snake_case ) ) def _lowercase ( self , _snake_case ) -> int: if right == 0: return 0 _UpperCamelCase : Dict = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] _UpperCamelCase : str = self.prev(_snake_case ) return result def _lowercase ( self , _snake_case , _snake_case ) -> int: return self.prefix(_snake_case ) - self.prefix(_snake_case ) def _lowercase ( self , _snake_case ) -> int: return self.query(_snake_case , index + 1 ) def _lowercase ( self , _snake_case ) -> int: value -= self.tree[0] if value < 0: return -1 _UpperCamelCase : Tuple = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 _UpperCamelCase : Optional[int] = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } _UpperCAmelCase : Any = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : Tuple = DistilBertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> int: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : int = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = do_lower_case _UpperCamelCase : Dict = strip_accents _UpperCamelCase : List[Any] = tokenize_chinese_chars _UpperCamelCase : Tuple = normalizer_class(**_snake_case ) _UpperCamelCase : Dict = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[int]: _UpperCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Optional[Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
1
'''simple docstring''' from math import pow, sqrt def snake_case__ ( *UpperCamelCase ) -> bool: _UpperCamelCase : Optional[Any] = len(UpperCamelCase ) > 0 and all(value > 0.0 for value in values ) return result def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(UpperCamelCase ,UpperCamelCase ) else ValueError('''Input Error: Molar mass values must greater than 0.''' ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) ,6 ) if validate(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a ,2 ) ,6 ) if validate(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a ,2 ) / molar_mass ,6 ) if validate(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else ValueError( '''Input Error: Molar mass and effusion rate values must greater than 0.''' ) )
683
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> list: _UpperCamelCase : Any = False while is_sorted is False: # Until all the indices are traversed keep looping _UpperCamelCase : List[str] = True for i in range(0 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Dict = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : int = False for i in range(1 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Optional[Any] = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : Optional[int] = False return input_list if __name__ == "__main__": print("""Enter list to be sorted""") _UpperCAmelCase : Optional[int] = [int(x) for x in input().split()] # inputing elements of the list in one line _UpperCAmelCase : Union[str, Any] = odd_even_sort(input_list) print("""The sorted list is""") print(sorted_list)
683
1
'''simple docstring''' import numpy as np def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase = 1e-12 ,UpperCamelCase = 1_00 ,) -> tuple[float, np.ndarray]: assert np.shape(UpperCamelCase )[0] == np.shape(UpperCamelCase )[1] # Ensure proper dimensionality. assert np.shape(UpperCamelCase )[0] == np.shape(UpperCamelCase )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(UpperCamelCase ) == np.iscomplexobj(UpperCamelCase ) _UpperCamelCase : Tuple = np.iscomplexobj(UpperCamelCase ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(UpperCamelCase ,input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _UpperCamelCase : Tuple = False _UpperCamelCase : Tuple = 0 _UpperCamelCase : Optional[int] = 0 _UpperCamelCase : List[str] = 1e12 while not convergence: # Multiple matrix by the vector. _UpperCamelCase : Optional[Any] = np.dot(UpperCamelCase ,UpperCamelCase ) # Normalize the resulting output vector. _UpperCamelCase : Optional[int] = w / np.linalg.norm(UpperCamelCase ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _UpperCamelCase : Union[str, Any] = vector.conj().T if is_complex else vector.T _UpperCamelCase : List[Any] = np.dot(UpperCamelCase ,np.dot(UpperCamelCase ,UpperCamelCase ) ) # Check convergence. _UpperCamelCase : Any = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _UpperCamelCase : Optional[int] = True _UpperCamelCase : Optional[int] = lambda_ if is_complex: _UpperCamelCase : str = np.real(lambda_ ) return lambda_, vector def snake_case__ ( ) -> None: _UpperCamelCase : Union[str, Any] = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _UpperCamelCase : List[Any] = np.array([41, 4, 20] ) _UpperCamelCase : Union[str, Any] = real_input_matrix.astype(np.complexaaa ) _UpperCamelCase : Any = np.triu(1j * complex_input_matrix ,1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _UpperCamelCase : List[str] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _UpperCamelCase : Optional[int] = real_input_matrix _UpperCamelCase : List[Any] = real_vector elif problem_type == "complex": _UpperCamelCase : Tuple = complex_input_matrix _UpperCamelCase : Tuple = complex_vector # Our implementation. _UpperCamelCase, _UpperCamelCase : Union[str, Any] = power_iteration(UpperCamelCase ,UpperCamelCase ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _UpperCamelCase, _UpperCamelCase : Union[str, Any] = np.linalg.eigh(UpperCamelCase ) # Last eigenvalue is the maximum one. _UpperCamelCase : int = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _UpperCamelCase : Optional[int] = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(UpperCamelCase ) - np.abs(UpperCamelCase ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
683
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = checkpoint _UpperCamelCase : int = {} _UpperCamelCase : int = vae_state_dict['''encoder.conv_in.weight'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_in.bias'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_out.weight'''] _UpperCamelCase : Any = vae_state_dict['''encoder.conv_out.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''encoder.norm_out.weight'''] _UpperCamelCase : str = vae_state_dict['''encoder.norm_out.bias'''] _UpperCamelCase : str = vae_state_dict['''decoder.conv_in.weight'''] _UpperCamelCase : List[Any] = vae_state_dict['''decoder.conv_in.bias'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.weight'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.bias'''] _UpperCamelCase : int = vae_state_dict['''decoder.norm_out.weight'''] _UpperCamelCase : Dict = vae_state_dict['''decoder.norm_out.bias'''] _UpperCamelCase : Optional[int] = vae_state_dict['''quant_conv.weight'''] _UpperCamelCase : int = vae_state_dict['''quant_conv.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''post_quant_conv.weight'''] _UpperCamelCase : Optional[int] = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only _UpperCamelCase : Optional[int] = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) _UpperCamelCase : Tuple = { layer_id: [key for key in vae_state_dict if f'''down.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } # Retrieves the keys for the decoder up blocks only _UpperCamelCase : Any = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) _UpperCamelCase : int = { layer_id: [key for key in vae_state_dict if f'''up.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } for i in range(UpperCamelCase ): _UpperCamelCase : Any = [key for key in down_blocks[i] if f'''down.{i}''' in key and f'''down.{i}.downsample''' not in key] if f'''encoder.down.{i}.downsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Optional[int] = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.weight''' ) _UpperCamelCase : Dict = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.bias''' ) _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Union[str, Any] = {'''old''': f'''down.{i}.block''', '''new''': f'''down_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[str] = [key for key in vae_state_dict if '''encoder.mid.block''' in key] _UpperCamelCase : Tuple = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : Optional[int] = [key for key in mid_resnets if f'''encoder.mid.block_{i}''' in key] _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Tuple = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] _UpperCamelCase : List[str] = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) for i in range(UpperCamelCase ): _UpperCamelCase : Union[str, Any] = num_up_blocks - 1 - i _UpperCamelCase : Optional[int] = [ key for key in up_blocks[block_id] if f'''up.{block_id}''' in key and f'''up.{block_id}.upsample''' not in key ] if f'''decoder.up.{block_id}.upsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Tuple = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.weight''' ] _UpperCamelCase : Any = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.bias''' ] _UpperCamelCase : Any = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': f'''up.{block_id}.block''', '''new''': f'''up_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[Any] = [key for key in vae_state_dict if '''decoder.mid.block''' in key] _UpperCamelCase : Optional[Any] = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : int = [key for key in mid_resnets if f'''decoder.mid.block_{i}''' in key] _UpperCamelCase : Optional[int] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Optional[Any] = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] _UpperCamelCase : Tuple = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Dict = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) return new_checkpoint def snake_case__ ( UpperCamelCase ,UpperCamelCase ,) -> List[str]: # Only support V1 _UpperCamelCase : Tuple = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) _UpperCamelCase : List[Any] = io.BytesIO(r.content ) _UpperCamelCase : Optional[int] = OmegaConf.load(UpperCamelCase ) _UpperCamelCase : str = 5_12 _UpperCamelCase : int = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open _UpperCamelCase : str = {} with safe_open(UpperCamelCase ,framework='''pt''' ,device='''cpu''' ) as f: for key in f.keys(): _UpperCamelCase : Union[str, Any] = f.get_tensor(UpperCamelCase ) else: _UpperCamelCase : str = torch.load(UpperCamelCase ,map_location=UpperCamelCase )['''state_dict'''] # Convert the VAE model. _UpperCamelCase : Dict = create_vae_diffusers_config(UpperCamelCase ,image_size=UpperCamelCase ) _UpperCamelCase : str = custom_convert_ldm_vae_checkpoint(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Dict = AutoencoderKL(**UpperCamelCase ) vae.load_state_dict(UpperCamelCase ) vae.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") _UpperCAmelCase : int = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
683
1
'''simple docstring''' from collections.abc import Sequence def snake_case__ ( UpperCamelCase ,UpperCamelCase = False ) -> float: if not arr: return 0 _UpperCamelCase : Optional[int] = 0 if allow_empty_subarrays else float('''-inf''' ) _UpperCamelCase : Any = 0.0 for num in arr: _UpperCamelCase : Tuple = max(0 if allow_empty_subarrays else num ,curr_sum + num ) _UpperCamelCase : Dict = max(UpperCamelCase ,UpperCamelCase ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() _UpperCAmelCase : str = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(f"""{max_subarray_sum(nums) = }""")
683
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = ['image_processor', 'tokenizer'] A__ : Dict = 'CLIPImageProcessor' A__ : str = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , _snake_case=None , _snake_case=None , **_snake_case ) -> List[Any]: _UpperCamelCase : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) _UpperCamelCase : Optional[Any] = kwargs.pop('''feature_extractor''' ) _UpperCamelCase : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_snake_case , _snake_case ) def __call__( self , _snake_case=None , _snake_case=None , _snake_case=None , **_snake_case ) -> Dict: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: _UpperCamelCase : List[str] = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: _UpperCamelCase : str = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: _UpperCamelCase : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Tuple: return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Any: return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def _lowercase ( self ) -> int: _UpperCamelCase : Optional[int] = self.tokenizer.model_input_names _UpperCamelCase : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
683
1
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase ( a_ ): """simple docstring""" A__ : UNetaDModel A__ : ScoreSdeVeScheduler def __init__( self , _snake_case , _snake_case ) -> Union[str, Any]: super().__init__() self.register_modules(unet=_snake_case , scheduler=_snake_case ) @torch.no_grad() def __call__( self , _snake_case = 1 , _snake_case = 2000 , _snake_case = None , _snake_case = "pil" , _snake_case = True , **_snake_case , ) -> Union[ImagePipelineOutput, Tuple]: _UpperCamelCase : str = self.unet.config.sample_size _UpperCamelCase : Optional[int] = (batch_size, 3, img_size, img_size) _UpperCamelCase : int = self.unet _UpperCamelCase : Optional[int] = randn_tensor(_snake_case , generator=_snake_case ) * self.scheduler.init_noise_sigma _UpperCamelCase : Union[str, Any] = sample.to(self.device ) self.scheduler.set_timesteps(_snake_case ) self.scheduler.set_sigmas(_snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): _UpperCamelCase : str = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): _UpperCamelCase : int = self.unet(_snake_case , _snake_case ).sample _UpperCamelCase : List[Any] = self.scheduler.step_correct(_snake_case , _snake_case , generator=_snake_case ).prev_sample # prediction step _UpperCamelCase : Optional[Any] = model(_snake_case , _snake_case ).sample _UpperCamelCase : Optional[int] = self.scheduler.step_pred(_snake_case , _snake_case , _snake_case , generator=_snake_case ) _UpperCamelCase, _UpperCamelCase : Optional[int] = output.prev_sample, output.prev_sample_mean _UpperCamelCase : int = sample_mean.clamp(0 , 1 ) _UpperCamelCase : Tuple = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCamelCase : Union[str, Any] = self.numpy_to_pil(_snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=_snake_case )
683
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _UpperCAmelCase : Union[str, Any] = (720, 1280) # Height, Width _UpperCAmelCase : str = (0.4, 0.6) # if height or width lower than this scale, drop it. _UpperCAmelCase : Optional[Any] = 1 / 100 _UpperCAmelCase : Optional[Any] = """""" _UpperCAmelCase : int = """""" _UpperCAmelCase : Union[str, Any] = """""" _UpperCAmelCase : List[Any] = 250 def snake_case__ ( ) -> None: _UpperCamelCase, _UpperCamelCase : List[Any] = get_dataset(UpperCamelCase ,UpperCamelCase ) for index in range(UpperCamelCase ): _UpperCamelCase : List[str] = random.sample(range(len(UpperCamelCase ) ) ,4 ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[str] = update_image_and_anno( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,filter_scale=UpperCamelCase ,) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _UpperCamelCase : List[str] = random_chars(32 ) _UpperCamelCase : List[str] = path.split(os.sep )[-1].rsplit('''.''' ,1 )[0] _UpperCamelCase : Any = f'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(f'''{file_root}.jpg''' ,UpperCamelCase ,[cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) _UpperCamelCase : Any = [] for anno in new_annos: _UpperCamelCase : List[Any] = anno[3] - anno[1] _UpperCamelCase : int = anno[4] - anno[2] _UpperCamelCase : int = anno[1] + width / 2 _UpperCamelCase : int = anno[2] + height / 2 _UpperCamelCase : Optional[Any] = f'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(UpperCamelCase ) with open(f'''{file_root}.txt''' ,'''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> tuple[list, list]: _UpperCamelCase : List[str] = [] _UpperCamelCase : Union[str, Any] = [] for label_file in glob.glob(os.path.join(UpperCamelCase ,'''*.txt''' ) ): _UpperCamelCase : int = label_file.split(os.sep )[-1].rsplit('''.''' ,1 )[0] with open(UpperCamelCase ) as in_file: _UpperCamelCase : Dict = in_file.readlines() _UpperCamelCase : Tuple = os.path.join(UpperCamelCase ,f'''{label_name}.jpg''' ) _UpperCamelCase : Tuple = [] for obj_list in obj_lists: _UpperCamelCase : List[Any] = obj_list.rstrip('''\n''' ).split(''' ''' ) _UpperCamelCase : Tuple = float(obj[1] ) - float(obj[3] ) / 2 _UpperCamelCase : Any = float(obj[2] ) - float(obj[4] ) / 2 _UpperCamelCase : Tuple = float(obj[1] ) + float(obj[3] ) / 2 _UpperCamelCase : List[Any] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(UpperCamelCase ) labels.append(UpperCamelCase ) return img_paths, labels def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = 0.0 ,) -> tuple[list, list, str]: _UpperCamelCase : Optional[int] = np.zeros([output_size[0], output_size[1], 3] ,dtype=np.uinta ) _UpperCamelCase : str = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = int(scale_x * output_size[1] ) _UpperCamelCase : Dict = int(scale_y * output_size[0] ) _UpperCamelCase : int = [] _UpperCamelCase : Union[str, Any] = [] for i, index in enumerate(UpperCamelCase ): _UpperCamelCase : Optional[int] = all_img_list[index] path_list.append(UpperCamelCase ) _UpperCamelCase : str = all_annos[index] _UpperCamelCase : Tuple = cva.imread(UpperCamelCase ) if i == 0: # top-left _UpperCamelCase : Any = cva.resize(UpperCamelCase ,(divid_point_x, divid_point_y) ) _UpperCamelCase : Any = img for bbox in img_annos: _UpperCamelCase : List[Any] = bbox[1] * scale_x _UpperCamelCase : Dict = bbox[2] * scale_y _UpperCamelCase : Any = bbox[3] * scale_x _UpperCamelCase : Any = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _UpperCamelCase : Union[str, Any] = cva.resize(UpperCamelCase ,(output_size[1] - divid_point_x, divid_point_y) ) _UpperCamelCase : List[Any] = img for bbox in img_annos: _UpperCamelCase : Any = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Optional[Any] = bbox[2] * scale_y _UpperCamelCase : Any = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : Optional[int] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _UpperCamelCase : Dict = cva.resize(UpperCamelCase ,(divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : List[str] = img for bbox in img_annos: _UpperCamelCase : int = bbox[1] * scale_x _UpperCamelCase : Optional[Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : int = bbox[3] * scale_x _UpperCamelCase : Any = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _UpperCamelCase : Dict = cva.resize( UpperCamelCase ,(output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : Union[str, Any] = img for bbox in img_annos: _UpperCamelCase : Optional[int] = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Union[str, Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : List[str] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _UpperCamelCase : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def snake_case__ ( UpperCamelCase ) -> str: assert number_char > 1, "The number of character should greater than 1" _UpperCamelCase : Tuple = ascii_lowercase + digits return "".join(random.choice(UpperCamelCase ) for _ in range(UpperCamelCase ) ) if __name__ == "__main__": main() print("""DONE ✅""")
683
1
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : List[str] = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} _UpperCAmelCase : int = { """vocab_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json""" ), }, """merges_file""": { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt""", """allenai/longformer-large-4096""": ( """https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt""" ), }, } _UpperCAmelCase : List[Any] = { """allenai/longformer-base-4096""": 4096, """allenai/longformer-large-4096""": 4096, """allenai/longformer-large-4096-finetuned-triviaqa""": 4096, """allenai/longformer-base-4096-extra.pos.embd.only""": 4096, """allenai/longformer-large-4096-extra.pos.embd.only""": 4096, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : int = ( list(range(ord('''!''' ) ,ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) ,ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) ,ord('''ÿ''' ) + 1 ) ) ) _UpperCamelCase : Optional[int] = bs[:] _UpperCamelCase : Optional[int] = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCamelCase ) cs.append(2**8 + n ) n += 1 _UpperCamelCase : Optional[Any] = [chr(UpperCamelCase ) for n in cs] return dict(zip(UpperCamelCase ,UpperCamelCase ) ) def snake_case__ ( UpperCamelCase ) -> Any: _UpperCamelCase : str = set() _UpperCamelCase : str = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _UpperCamelCase : Optional[Any] = char return pairs class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[int] = ['input_ids', 'attention_mask'] def __init__( self , _snake_case , _snake_case , _snake_case="replace" , _snake_case="<s>" , _snake_case="</s>" , _snake_case="</s>" , _snake_case="<s>" , _snake_case="<unk>" , _snake_case="<pad>" , _snake_case="<mask>" , _snake_case=False , **_snake_case , ) -> Dict: _UpperCamelCase : int = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else bos_token _UpperCamelCase : int = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else eos_token _UpperCamelCase : List[Any] = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else sep_token _UpperCamelCase : Any = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else cls_token _UpperCamelCase : int = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else unk_token _UpperCamelCase : List[Any] = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _UpperCamelCase : Any = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else mask_token super().__init__( errors=_snake_case , bos_token=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , cls_token=_snake_case , pad_token=_snake_case , mask_token=_snake_case , add_prefix_space=_snake_case , **_snake_case , ) with open(_snake_case , encoding='''utf-8''' ) as vocab_handle: _UpperCamelCase : Optional[Any] = json.load(_snake_case ) _UpperCamelCase : Union[str, Any] = {v: k for k, v in self.encoder.items()} _UpperCamelCase : Any = errors # how to handle errors in decoding _UpperCamelCase : Optional[int] = bytes_to_unicode() _UpperCamelCase : Union[str, Any] = {v: k for k, v in self.byte_encoder.items()} with open(_snake_case , encoding='''utf-8''' ) as merges_handle: _UpperCamelCase : List[str] = merges_handle.read().split('''\n''' )[1:-1] _UpperCamelCase : Any = [tuple(merge.split() ) for merge in bpe_merges] _UpperCamelCase : Optional[int] = dict(zip(_snake_case , range(len(_snake_case ) ) ) ) _UpperCamelCase : Any = {} _UpperCamelCase : Any = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _UpperCamelCase : int = re.compile(r'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property def _lowercase ( self ) -> Tuple: return len(self.encoder ) def _lowercase ( self ) -> Union[str, Any]: return dict(self.encoder , **self.added_tokens_encoder ) def _lowercase ( self , _snake_case ) -> Union[str, Any]: if token in self.cache: return self.cache[token] _UpperCamelCase : Dict = tuple(_snake_case ) _UpperCamelCase : List[Any] = get_pairs(_snake_case ) if not pairs: return token while True: _UpperCamelCase : str = min(_snake_case , key=lambda _snake_case : self.bpe_ranks.get(_snake_case , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _UpperCamelCase, _UpperCamelCase : Optional[int] = bigram _UpperCamelCase : Optional[Any] = [] _UpperCamelCase : Any = 0 while i < len(_snake_case ): try: _UpperCamelCase : Tuple = word.index(_snake_case , _snake_case ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _UpperCamelCase : Optional[int] = j if word[i] == first and i < len(_snake_case ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _UpperCamelCase : Optional[Any] = tuple(_snake_case ) _UpperCamelCase : Optional[int] = new_word if len(_snake_case ) == 1: break else: _UpperCamelCase : Dict = get_pairs(_snake_case ) _UpperCamelCase : Tuple = ''' '''.join(_snake_case ) _UpperCamelCase : Optional[Any] = word return word def _lowercase ( self , _snake_case ) -> int: _UpperCamelCase : Dict = [] for token in re.findall(self.pat , _snake_case ): _UpperCamelCase : int = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_snake_case ).split(''' ''' ) ) return bpe_tokens def _lowercase ( self , _snake_case ) -> Dict: return self.encoder.get(_snake_case , self.encoder.get(self.unk_token ) ) def _lowercase ( self , _snake_case ) -> str: return self.decoder.get(_snake_case ) def _lowercase ( self , _snake_case ) -> List[str]: _UpperCamelCase : Union[str, Any] = ''''''.join(_snake_case ) _UpperCamelCase : List[Any] = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: if not os.path.isdir(_snake_case ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase : List[str] = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCamelCase : int = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(_snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_snake_case , ensure_ascii=_snake_case ) + '''\n''' ) _UpperCamelCase : Any = 0 with open(_snake_case , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _snake_case : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' ''' Please check that the tokenizer is not corrupted!''' ) _UpperCamelCase : List[str] = token_index writer.write(''' '''.join(_snake_case ) + '''\n''' ) index += 1 return vocab_file, merge_file def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCamelCase : List[str] = [self.cls_token_id] _UpperCamelCase : List[str] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowercase ( self , _snake_case , _snake_case = None , _snake_case = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_snake_case , token_ids_a=_snake_case , already_has_special_tokens=_snake_case ) if token_ids_a is None: return [1] + ([0] * len(_snake_case )) + [1] return [1] + ([0] * len(_snake_case )) + [1, 1] + ([0] * len(_snake_case )) + [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Tuple = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowercase ( self , _snake_case , _snake_case=False , **_snake_case ) -> Any: _UpperCamelCase : List[Any] = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_snake_case ) > 0 and not text[0].isspace()): _UpperCamelCase : List[str] = ''' ''' + text return (text, kwargs)
683
'''simple docstring''' from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class UpperCAmelCase ( a_ ): """simple docstring""" @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Union[str, Any] = EncoderDecoderModel.from_encoder_decoder_pretrained('''prajjwal1/bert-tiny''' , '''prajjwal1/bert-tiny''' ) _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained('''bert-base-uncased''' ) _UpperCamelCase : Optional[Any] = bertabert.config.encoder.vocab_size _UpperCamelCase : List[str] = tokenizer.sep_token_id _UpperCamelCase : List[str] = tokenizer.cls_token_id _UpperCamelCase : Optional[Any] = 128 _UpperCamelCase : int = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''train[:1%]''' ) _UpperCamelCase : Dict = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''validation[:1%]''' ) _UpperCamelCase : Dict = train_dataset.select(range(32 ) ) _UpperCamelCase : Tuple = val_dataset.select(range(16 ) ) _UpperCamelCase : Union[str, Any] = 4 def _map_to_encoder_decoder_inputs(_snake_case ): # Tokenizer will automatically set [BOS] <text> [EOS] _UpperCamelCase : Optional[Any] = tokenizer(batch['''article'''] , padding='''max_length''' , truncation=_snake_case , max_length=512 ) _UpperCamelCase : Optional[int] = tokenizer(batch['''highlights'''] , padding='''max_length''' , truncation=_snake_case , max_length=128 ) _UpperCamelCase : str = inputs.input_ids _UpperCamelCase : Union[str, Any] = inputs.attention_mask _UpperCamelCase : str = outputs.input_ids _UpperCamelCase : str = outputs.input_ids.copy() _UpperCamelCase : Tuple = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['''labels'''] ] _UpperCamelCase : Union[str, Any] = outputs.attention_mask assert all(len(_snake_case ) == 512 for x in inputs.input_ids ) assert all(len(_snake_case ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_snake_case ): _UpperCamelCase : Dict = pred.label_ids _UpperCamelCase : Optional[int] = pred.predictions # all unnecessary tokens are removed _UpperCamelCase : Any = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : Dict = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_snake_case ) )] ) / len(_snake_case ) return {"accuracy": accuracy} # map train dataset _UpperCamelCase : Optional[Any] = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) train_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) # same for validation dataset _UpperCamelCase : List[Any] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) val_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) _UpperCamelCase : Optional[int] = self.get_auto_remove_tmp_dir() _UpperCamelCase : Union[str, Any] = SeqaSeqTrainingArguments( output_dir=_snake_case , per_device_train_batch_size=_snake_case , per_device_eval_batch_size=_snake_case , predict_with_generate=_snake_case , evaluation_strategy='''steps''' , do_train=_snake_case , do_eval=_snake_case , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _UpperCamelCase : Optional[int] = SeqaSeqTrainer( model=_snake_case , args=_snake_case , compute_metrics=_compute_metrics , train_dataset=_snake_case , eval_dataset=_snake_case , tokenizer=_snake_case , ) # start training trainer.train()
683
1
'''simple docstring''' from __future__ import annotations from functools import lru_cache from math import ceil _UpperCAmelCase : int = 100 _UpperCAmelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _UpperCAmelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_00 ) def snake_case__ ( UpperCamelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _UpperCamelCase : set[int] = set() _UpperCamelCase : int _UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def snake_case__ ( UpperCamelCase = 50_00 ) -> int | None: for number_to_partition in range(1 ,UpperCamelCase ): if len(partition(UpperCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
683
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def snake_case__ ( UpperCamelCase=None ) -> Optional[int]: if subparsers is not None: _UpperCamelCase : Dict = subparsers.add_parser('''env''' ) else: _UpperCamelCase : Tuple = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''' ,default=UpperCamelCase ,help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase ) return parser def snake_case__ ( UpperCamelCase ) -> Any: _UpperCamelCase : int = torch.__version__ _UpperCamelCase : int = torch.cuda.is_available() _UpperCamelCase : List[str] = is_xpu_available() _UpperCamelCase : Dict = is_npu_available() _UpperCamelCase : Optional[Any] = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(UpperCamelCase ): _UpperCamelCase : List[str] = load_config_from_file(args.config_file ).to_dict() _UpperCamelCase : List[Any] = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': f'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(UpperCamelCase ), '''PyTorch NPU available''': str(UpperCamelCase ), '''System RAM''': f'''{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB''', } if pt_cuda_available: _UpperCamelCase : int = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([f'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) _UpperCamelCase : Union[str, Any] = ( '''\n'''.join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(UpperCamelCase ,UpperCamelCase ) else f'''\t{accelerate_config}''' ) print(UpperCamelCase ) _UpperCamelCase : str = accelerate_config return info def snake_case__ ( ) -> int: _UpperCamelCase : str = env_command_parser() _UpperCamelCase : Any = parser.parse_args() env_command(UpperCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
683
1
'''simple docstring''' from __future__ import annotations from collections import namedtuple def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> tuple: _UpperCamelCase : Tuple = namedtuple('''result''' ,'''name value''' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('''Only one argument must be 0''' ) elif power < 0: raise ValueError( '''Power cannot be negative in any electrical/electronics system''' ) elif voltage == 0: return result('''voltage''' ,power / current ) elif current == 0: return result('''current''' ,power / voltage ) elif power == 0: return result('''power''' ,float(round(abs(voltage * current ) ,2 ) ) ) else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : List[Any] = logging.get_logger(__name__) def snake_case__ ( UpperCamelCase ) -> Tuple: _UpperCamelCase : str = '''huggingface/label-files''' _UpperCamelCase : Optional[Any] = '''imagenet-1k-id2label.json''' _UpperCamelCase : Optional[int] = json.load(open(hf_hub_download(UpperCamelCase ,UpperCamelCase ,repo_type='''dataset''' ) ,'''r''' ) ) _UpperCamelCase : Optional[int] = {int(UpperCamelCase ): v for k, v in idalabel.items()} _UpperCamelCase : Dict = {v: k for k, v in idalabel.items()} _UpperCamelCase : Optional[Any] = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" _UpperCamelCase : Union[str, Any] = BitConfig( conv_layer=UpperCamelCase ,num_labels=10_00 ,idalabel=UpperCamelCase ,labelaid=UpperCamelCase ,) return config def snake_case__ ( UpperCamelCase ) -> str: if "stem.conv" in name: _UpperCamelCase : Any = name.replace('''stem.conv''' ,'''bit.embedder.convolution''' ) if "blocks" in name: _UpperCamelCase : Union[str, Any] = name.replace('''blocks''' ,'''layers''' ) if "head.fc" in name: _UpperCamelCase : Optional[Any] = name.replace('''head.fc''' ,'''classifier.1''' ) if name.startswith('''norm''' ): _UpperCamelCase : Any = '''bit.''' + name if "bit" not in name and "classifier" not in name: _UpperCamelCase : List[Any] = '''bit.encoder.''' + name return name def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase : List[str] = Image.open(requests.get(UpperCamelCase ,stream=UpperCamelCase ).raw ) return im @torch.no_grad() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[Any]: _UpperCamelCase : str = get_config(UpperCamelCase ) # load original model from timm _UpperCamelCase : int = create_model(UpperCamelCase ,pretrained=UpperCamelCase ) timm_model.eval() # load state_dict of original model _UpperCamelCase : int = timm_model.state_dict() for key in state_dict.copy().keys(): _UpperCamelCase : int = state_dict.pop(UpperCamelCase ) _UpperCamelCase : Any = val.squeeze() if '''head''' in key else val # load HuggingFace model _UpperCamelCase : List[str] = BitForImageClassification(UpperCamelCase ) model.eval() model.load_state_dict(UpperCamelCase ) # create image processor _UpperCamelCase : Optional[int] = create_transform(**resolve_data_config({} ,model=UpperCamelCase ) ) _UpperCamelCase : Any = transform.transforms _UpperCamelCase : List[str] = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } _UpperCamelCase : List[str] = BitImageProcessor( do_resize=UpperCamelCase ,size={'''shortest_edge''': timm_transforms[0].size} ,resample=pillow_resamplings[timm_transforms[0].interpolation.value] ,do_center_crop=UpperCamelCase ,crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} ,do_normalize=UpperCamelCase ,image_mean=timm_transforms[-1].mean.tolist() ,image_std=timm_transforms[-1].std.tolist() ,) _UpperCamelCase : str = prepare_img() _UpperCamelCase : Dict = transform(UpperCamelCase ).unsqueeze(0 ) _UpperCamelCase : Dict = processor(UpperCamelCase ,return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(UpperCamelCase ,UpperCamelCase ) # verify logits with torch.no_grad(): _UpperCamelCase : Optional[int] = model(UpperCamelCase ) _UpperCamelCase : Optional[int] = outputs.logits print('''Logits:''' ,logits[0, :3] ) print('''Predicted class:''' ,model.config.idalabel[logits.argmax(-1 ).item()] ) _UpperCamelCase : List[Any] = timm_model(UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase ,outputs.logits ,atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": _UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) _UpperCAmelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
683
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = 1 _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Tuple = (32, 32) _UpperCamelCase : int = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_snake_case ) return image @property def _lowercase ( self ) -> Optional[Any]: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def _lowercase ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def _lowercase ( self ) -> int: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(_snake_case ) @property def _lowercase ( self ) -> Tuple: def extract(*_snake_case , **_snake_case ): class UpperCAmelCase : """simple docstring""" def __init__( self ) -> Optional[Any]: _UpperCamelCase : Any = torch.ones([0] ) def _lowercase ( self , _snake_case ) -> str: self.pixel_values.to(_snake_case ) return self return Out() return extract def _lowercase ( self ) -> Any: _UpperCamelCase : str = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : List[str] = self.dummy_cond_unet _UpperCamelCase : List[Any] = PNDMScheduler(skip_prk_steps=_snake_case ) _UpperCamelCase : str = self.dummy_vae _UpperCamelCase : List[Any] = self.dummy_text_encoder _UpperCamelCase : Any = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) _UpperCamelCase : Dict = 77 _UpperCamelCase : Any = self.dummy_image.to(_snake_case ) _UpperCamelCase : List[str] = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk _UpperCamelCase : Optional[Any] = AltDiffusionImgaImgPipeline( unet=_snake_case , scheduler=_snake_case , vae=_snake_case , text_encoder=_snake_case , tokenizer=_snake_case , safety_checker=_snake_case , feature_extractor=self.dummy_extractor , ) _UpperCamelCase : str = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=_snake_case ) _UpperCamelCase : Optional[int] = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : List[Any] = '''A painting of a squirrel eating a burger''' _UpperCamelCase : List[str] = torch.Generator(device=_snake_case ).manual_seed(0 ) _UpperCamelCase : List[str] = alt_pipe( [prompt] , generator=_snake_case , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=_snake_case , ) _UpperCamelCase : Any = output.images _UpperCamelCase : List[str] = torch.Generator(device=_snake_case ).manual_seed(0 ) _UpperCamelCase : Optional[Any] = alt_pipe( [prompt] , generator=_snake_case , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=_snake_case , return_dict=_snake_case , )[0] _UpperCamelCase : Tuple = image[0, -3:, -3:, -1] _UpperCamelCase : List[str] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _UpperCamelCase : int = np.array([0.4_427, 0.3_731, 0.4_249, 0.4_941, 0.4_546, 0.4_148, 0.4_193, 0.4_666, 0.4_499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = self.dummy_cond_unet _UpperCamelCase : Dict = PNDMScheduler(skip_prk_steps=_snake_case ) _UpperCamelCase : Union[str, Any] = self.dummy_vae _UpperCamelCase : Dict = self.dummy_text_encoder _UpperCamelCase : Tuple = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) _UpperCamelCase : Optional[int] = 77 _UpperCamelCase : List[Any] = self.dummy_image.to(_snake_case ) # put models in fp16 _UpperCamelCase : Optional[int] = unet.half() _UpperCamelCase : Optional[Any] = vae.half() _UpperCamelCase : Tuple = bert.half() # make sure here that pndm scheduler skips prk _UpperCamelCase : Optional[Any] = AltDiffusionImgaImgPipeline( unet=_snake_case , scheduler=_snake_case , vae=_snake_case , text_encoder=_snake_case , tokenizer=_snake_case , safety_checker=_snake_case , feature_extractor=self.dummy_extractor , ) _UpperCamelCase : List[Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=_snake_case ) _UpperCamelCase : int = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : int = '''A painting of a squirrel eating a burger''' _UpperCamelCase : int = torch.manual_seed(0 ) _UpperCamelCase : Any = alt_pipe( [prompt] , generator=_snake_case , num_inference_steps=2 , output_type='''np''' , image=_snake_case , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : List[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) # resize to resolution that is divisible by 8 but not 16 or 32 _UpperCamelCase : Any = init_image.resize((760, 504) ) _UpperCamelCase : int = '''BAAI/AltDiffusion''' _UpperCamelCase : str = AltDiffusionImgaImgPipeline.from_pretrained( _snake_case , safety_checker=_snake_case , ) pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) pipe.enable_attention_slicing() _UpperCamelCase : List[str] = '''A fantasy landscape, trending on artstation''' _UpperCamelCase : Optional[Any] = torch.manual_seed(0 ) _UpperCamelCase : str = pipe( prompt=_snake_case , image=_snake_case , strength=0.75 , guidance_scale=7.5 , generator=_snake_case , output_type='''np''' , ) _UpperCamelCase : Dict = output.images[0] _UpperCamelCase : int = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) _UpperCamelCase : str = np.array([0.9_358, 0.9_397, 0.9_599, 0.9_901, 1.0_000, 1.0_000, 0.9_882, 1.0_000, 1.0_000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) _UpperCamelCase : Tuple = init_image.resize((768, 512) ) _UpperCamelCase : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''' ) _UpperCamelCase : Optional[int] = '''BAAI/AltDiffusion''' _UpperCamelCase : Optional[int] = AltDiffusionImgaImgPipeline.from_pretrained( _snake_case , safety_checker=_snake_case , ) pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) pipe.enable_attention_slicing() _UpperCamelCase : Dict = '''A fantasy landscape, trending on artstation''' _UpperCamelCase : List[str] = torch.manual_seed(0 ) _UpperCamelCase : List[Any] = pipe( prompt=_snake_case , image=_snake_case , strength=0.75 , guidance_scale=7.5 , generator=_snake_case , output_type='''np''' , ) _UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
683
'''simple docstring''' _UpperCAmelCase : Any = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100000)] def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : Any = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_00_00] number //= 10_00_00 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _UpperCAmelCase : list[bool | None] = [None] * 10000000 _UpperCAmelCase : str = True _UpperCAmelCase : Tuple = False def snake_case__ ( UpperCamelCase ) -> bool: if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _UpperCamelCase : List[str] = chain(next_number(UpperCamelCase ) ) _UpperCamelCase : Tuple = number_chain while number < 10_00_00_00: _UpperCamelCase : int = number_chain number *= 10 return number_chain def snake_case__ ( UpperCamelCase = 10_00_00_00 ) -> int: for i in range(1 ,UpperCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
683
1
'''simple docstring''' from manim import * class UpperCAmelCase ( a_ ): """simple docstring""" def _lowercase ( self ) -> str: _UpperCamelCase : str = Rectangle(height=0.5 , width=0.5 ) _UpperCamelCase : int = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _UpperCamelCase : int = Rectangle(height=0.25 , width=0.25 ) _UpperCamelCase : Any = [mem.copy() for i in range(6 )] _UpperCamelCase : str = [mem.copy() for i in range(6 )] _UpperCamelCase : str = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : Union[str, Any] = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : Tuple = VGroup(_snake_case , _snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : List[Any] = Text('''CPU''' , font_size=24 ) _UpperCamelCase : Optional[Any] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_snake_case ) _UpperCamelCase : int = [mem.copy() for i in range(4 )] _UpperCamelCase : Optional[int] = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : List[str] = Text('''GPU''' , font_size=24 ) _UpperCamelCase : List[str] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) gpu.move_to([-1, -1, 0] ) self.add(_snake_case ) _UpperCamelCase : Dict = [mem.copy() for i in range(6 )] _UpperCamelCase : List[Any] = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : Optional[Any] = Text('''Model''' , font_size=24 ) _UpperCamelCase : List[Any] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) model.move_to([3, -1.0, 0] ) self.add(_snake_case ) _UpperCamelCase : Tuple = [] _UpperCamelCase : Optional[Any] = [] for i, rect in enumerate(_snake_case ): _UpperCamelCase : Optional[int] = fill.copy().set_fill(_snake_case , opacity=0.8 ) target.move_to(_snake_case ) model_arr.append(_snake_case ) _UpperCamelCase : List[str] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(_snake_case , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(_snake_case ) self.add(*_snake_case , *_snake_case ) _UpperCamelCase : Dict = [meta_mem.copy() for i in range(6 )] _UpperCamelCase : Optional[Any] = [meta_mem.copy() for i in range(6 )] _UpperCamelCase : Any = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : List[str] = VGroup(*_snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : Union[str, Any] = VGroup(_snake_case , _snake_case ).arrange(_snake_case , buff=0 ) _UpperCamelCase : List[str] = Text('''Disk''' , font_size=24 ) _UpperCamelCase : List[Any] = Group(_snake_case , _snake_case ).arrange(_snake_case , buff=0.5 , aligned_edge=_snake_case ) disk.move_to([-4, -1.25, 0] ) self.add(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCamelCase : Optional[Any] = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_snake_case ) _UpperCamelCase : List[str] = MarkupText( F'''Now watch as an input is passed through the model\nand how the memory is utilized and handled.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case ) ) _UpperCamelCase : Optional[Any] = Square(0.3 ) input.set_fill(_snake_case , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , _snake_case , buff=0.5 ) self.play(Write(_snake_case ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=_snake_case , buff=0.02 ) self.play(MoveToTarget(_snake_case ) ) self.play(FadeOut(_snake_case ) ) _UpperCamelCase : Optional[Any] = Arrow(start=_snake_case , end=_snake_case , color=_snake_case , buff=0.5 ) a.next_to(model_arr[0].get_left() , _snake_case , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) _UpperCamelCase : Optional[Any] = MarkupText( F'''As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case , run_time=3 ) ) _UpperCamelCase : List[Any] = {'''run_time''': 1, '''fade_in''': True, '''fade_out''': True, '''buff''': 0.02} self.play( Write(_snake_case ) , Circumscribe(model_arr[0] , color=_snake_case , **_snake_case ) , Circumscribe(model_cpu_arr[0] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) _UpperCamelCase : Optional[Any] = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , _snake_case , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) _UpperCamelCase : Tuple = AnimationGroup( FadeOut(_snake_case , run_time=0.5 ) , MoveToTarget(_snake_case , run_time=0.5 ) , FadeIn(_snake_case , run_time=0.5 ) , lag_ratio=0.2 ) self.play(_snake_case ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: _UpperCamelCase : Any = 0.7 self.play( Circumscribe(model_arr[i] , **_snake_case ) , Circumscribe(cpu_left_col_base[i] , **_snake_case ) , Circumscribe(cpu_left_col_base[i + 1] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , Circumscribe(model_arr[i + 1] , color=_snake_case , **_snake_case ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=_snake_case , **_snake_case ) , Circumscribe(cpu_left_col_base[-1] , color=_snake_case , **_snake_case ) , Circumscribe(gpu_rect[0] , color=_snake_case , **_snake_case ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) _UpperCamelCase : Dict = a_c _UpperCamelCase : Optional[int] = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(_snake_case ) , FadeOut(_snake_case , run_time=0.5 ) , ) _UpperCamelCase : List[str] = MarkupText(F'''Inference on a model too large for GPU memory\nis successfully completed.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_snake_case , run_time=3 ) , MoveToTarget(_snake_case ) ) self.wait()
683
'''simple docstring''' _UpperCAmelCase : str = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : List[str] = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> str: assert len(str(UpperCamelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _UpperCamelCase : Any = year // 1_00 _UpperCamelCase : List[Any] = (5 * (century % 4) + 2) % 7 _UpperCamelCase : Tuple = year % 1_00 _UpperCamelCase : Optional[int] = centurian % 12 _UpperCamelCase : Tuple = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _UpperCamelCase : List[Any] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) _UpperCamelCase : Optional[int] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
683
1
'''simple docstring''' from typing import Dict, Optional import numpy as np import datasets _UpperCAmelCase : Dict = """ IoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union between the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation, the mean IoU of the image is calculated by taking the IoU of each class and averaging them. """ _UpperCAmelCase : Dict = """ Args: predictions (`List[ndarray]`): List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. references (`List[ndarray]`): List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. num_labels (`int`): Number of classes (categories). ignore_index (`int`): Index that will be ignored during evaluation. nan_to_num (`int`, *optional*): If specified, NaN values will be replaced by the number defined by the user. label_map (`dict`, *optional*): If specified, dictionary mapping old label indices to new label indices. reduce_labels (`bool`, *optional*, defaults to `False`): Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background, and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255. Returns: `Dict[str, float | ndarray]` comprising various elements: - *mean_iou* (`float`): Mean Intersection-over-Union (IoU averaged over all categories). - *mean_accuracy* (`float`): Mean accuracy (averaged over all categories). - *overall_accuracy* (`float`): Overall accuracy on all images. - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`): Per category accuracy. - *per_category_iou* (`ndarray` of shape `(num_labels,)`): Per category IoU. Examples: >>> import numpy as np >>> mean_iou = datasets.load_metric(\"mean_iou\") >>> # suppose one has 3 different segmentation maps predicted >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]]) >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]]) >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]]) >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]]) >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]]) >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]]) >>> predicted = [predicted_1, predicted_2, predicted_3] >>> ground_truth = [actual_1, actual_2, actual_3] >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False) >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mean_iou': 0.47750000000000004, 'mean_accuracy': 0.5916666666666666, 'overall_accuracy': 0.5263157894736842, 'per_category_iou': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), 'per_category_accuracy': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])} """ _UpperCAmelCase : str = """\ @software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020, author = {{MMSegmentation Contributors}}, license = {Apache-2.0}, month = {7}, title = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}}, url = {https://github.com/open-mmlab/mmsegmentation}, year = {2020} }""" def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = None ,UpperCamelCase = False ,) -> int: if label_map is not None: for old_id, new_id in label_map.items(): _UpperCamelCase : Optional[int] = new_id # turn into Numpy arrays _UpperCamelCase : Any = np.array(UpperCamelCase ) _UpperCamelCase : Tuple = np.array(UpperCamelCase ) if reduce_labels: _UpperCamelCase : Optional[int] = 2_55 _UpperCamelCase : Optional[Any] = label - 1 _UpperCamelCase : Optional[int] = 2_55 _UpperCamelCase : Union[str, Any] = label != ignore_index _UpperCamelCase : List[Any] = np.not_equal(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Dict = pred_label[mask] _UpperCamelCase : List[Any] = np.array(UpperCamelCase )[mask] _UpperCamelCase : Any = pred_label[pred_label == label] _UpperCamelCase : List[Any] = np.histogram(UpperCamelCase ,bins=UpperCamelCase ,range=(0, num_labels - 1) )[0] _UpperCamelCase : int = np.histogram(UpperCamelCase ,bins=UpperCamelCase ,range=(0, num_labels - 1) )[0] _UpperCamelCase : Optional[int] = np.histogram(UpperCamelCase ,bins=UpperCamelCase ,range=(0, num_labels - 1) )[0] _UpperCamelCase : Optional[int] = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = None ,UpperCamelCase = False ,) -> Tuple: _UpperCamelCase : int = np.zeros((num_labels,) ,dtype=np.floataa ) _UpperCamelCase : List[str] = np.zeros((num_labels,) ,dtype=np.floataa ) _UpperCamelCase : List[Any] = np.zeros((num_labels,) ,dtype=np.floataa ) _UpperCamelCase : int = np.zeros((num_labels,) ,dtype=np.floataa ) for result, gt_seg_map in zip(UpperCamelCase ,UpperCamelCase ): _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = intersect_and_union( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = None ,UpperCamelCase = None ,UpperCamelCase = False ,) -> Tuple: _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : int = total_intersect_and_union( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # compute metrics _UpperCamelCase : List[str] = {} _UpperCamelCase : str = total_area_intersect.sum() / total_area_label.sum() _UpperCamelCase : str = total_area_intersect / total_area_union _UpperCamelCase : str = total_area_intersect / total_area_label _UpperCamelCase : Dict = np.nanmean(UpperCamelCase ) _UpperCamelCase : Dict = np.nanmean(UpperCamelCase ) _UpperCamelCase : List[Any] = all_acc _UpperCamelCase : int = iou _UpperCamelCase : Optional[int] = acc if nan_to_num is not None: _UpperCamelCase : List[Any] = {metric: np.nan_to_num(UpperCamelCase ,nan=UpperCamelCase ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): """simple docstring""" def _lowercase ( self ) -> List[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { '''predictions''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), '''references''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), } ) , reference_urls=[ '''https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py''' ] , ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = None , _snake_case = None , _snake_case = False , ) -> str: _UpperCamelCase : Optional[int] = mean_iou( results=_snake_case , gt_seg_maps=_snake_case , num_labels=_snake_case , ignore_index=_snake_case , nan_to_num=_snake_case , label_map=_snake_case , reduce_labels=_snake_case , ) return iou_result
683
'''simple docstring''' import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : """simple docstring""" @staticmethod def _lowercase ( *_snake_case , **_snake_case ) -> str: pass @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" A__ : Tuple = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Optional[Any]: _UpperCamelCase : int = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Any = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def _lowercase ( self , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : int = vqa_pipeline(_snake_case , top_k=1 ) self.assertEqual( _snake_case , [ [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], ] , ) @require_torch def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[int] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question='''How many cats are there?''' , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) _UpperCamelCase : List[Any] = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[Any] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question=_snake_case , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : str = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : Dict = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [[{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''' ) def _lowercase ( self ) -> List[Any]: pass
683
1
'''simple docstring''' import os import jsonlines import numpy as np from tqdm import tqdm _UpperCAmelCase : Tuple = 2048 _UpperCAmelCase : int = 4096 _UpperCAmelCase : Dict = 42 _UpperCAmelCase : str = os.environ.pop("""PROCESS_TRAIN""", """false""") _UpperCAmelCase : List[str] = {"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def snake_case__ ( UpperCamelCase ) -> Union[str, Any]: def choose_first(UpperCamelCase ,UpperCamelCase=False ): assert isinstance(UpperCamelCase ,UpperCamelCase ) if len(UpperCamelCase ) == 1: _UpperCamelCase : Optional[int] = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: _UpperCamelCase : Union[str, Any] = {k: [a[k]] for k in a} if len(a['''start_token'''] ) > 0: break return a _UpperCamelCase : int = {'''id''': example['''id''']} _UpperCamelCase : str = example['''annotations'''] _UpperCamelCase : Optional[Any] = annotation['''yes_no_answer'''] if 0 in yes_no_answer or 1 in yes_no_answer: _UpperCamelCase : List[Any] = ['''yes'''] if 1 in yes_no_answer else ['''no'''] _UpperCamelCase : List[str] = [] _UpperCamelCase : Dict = [] _UpperCamelCase : Optional[Any] = ['''<cls>'''] else: _UpperCamelCase : str = ['''short'''] _UpperCamelCase : int = choose_first(annotation['''short_answers'''] ) if len(out['''start_token'''] ) == 0: # answer will be long if short is not available _UpperCamelCase : str = ['''long'''] _UpperCamelCase : Optional[int] = choose_first(annotation['''long_answer'''] ,is_long_answer=UpperCamelCase ) _UpperCamelCase : Optional[int] = [] answer.update(UpperCamelCase ) # disregard some samples if len(answer['''start_token'''] ) > 1 or answer["start_token"] == answer["end_token"]: _UpperCamelCase : Tuple = True else: _UpperCamelCase : Any = False _UpperCamelCase : Dict = ['''start_token''', '''end_token''', '''start_byte''', '''end_byte''', '''text'''] if not all(isinstance(answer[k] ,UpperCamelCase ) for k in cols ): raise ValueError('''Issue in ID''' ,example['''id'''] ) return answer def snake_case__ ( UpperCamelCase ,UpperCamelCase=False ) -> List[str]: _UpperCamelCase : Union[str, Any] = _get_single_answer(UpperCamelCase ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element _UpperCamelCase : Tuple = example['''document''']['''tokens'''] _UpperCamelCase : List[str] = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) return { "context": " ".join(UpperCamelCase ), "answer": { "start_token": -1_00, # ignore index in cross-entropy "end_token": -1_00, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples _UpperCamelCase : str = ['''start_token''', '''end_token'''] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 _UpperCamelCase : List[str] = example['''document''']['''tokens'''] _UpperCamelCase : int = answer['''start_token'''] _UpperCamelCase : List[str] = answer['''end_token'''] _UpperCamelCase : Optional[Any] = [] for i in range(len(doc['''token'''] ) ): if not doc["is_html"][i]: context.append(doc['''token'''][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 _UpperCamelCase : List[Any] = ''' '''.join(context[start_token:end_token] ) # checking above code if assertion: _UpperCamelCase : Optional[int] = doc['''is_html'''][answer['''start_token'''] : answer['''end_token''']] _UpperCamelCase : Union[str, Any] = doc['''token'''][answer['''start_token'''] : answer['''end_token''']] _UpperCamelCase : Tuple = ''' '''.join([old[i] for i in range(len(UpperCamelCase ) ) if not is_html[i]] ) if new != old: print('''ID:''' ,example['''id'''] ) print('''New:''' ,UpperCamelCase ,end='''\n''' ) print('''Old:''' ,UpperCamelCase ,end='''\n\n''' ) return { "context": " ".join(UpperCamelCase ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase=20_48 ,UpperCamelCase=40_96 ,UpperCamelCase=True ) -> Any: # overlap will be of doc_stride - q_len _UpperCamelCase : Optional[Any] = get_context_and_ans(UpperCamelCase ,assertion=UpperCamelCase ) _UpperCamelCase : Tuple = out['''answer'''] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } _UpperCamelCase : str = tokenizer(example['''question''']['''text'''] ,out['''context'''] ).input_ids _UpperCamelCase : Dict = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element _UpperCamelCase : Union[str, Any] = [] _UpperCamelCase : str = [] _UpperCamelCase : int = input_ids[:q_len] _UpperCamelCase : Optional[int] = range(UpperCamelCase ,len(UpperCamelCase ) ,max_length - doc_stride ) for i in doc_start_indices: _UpperCamelCase : List[str] = i + max_length - q_len _UpperCamelCase : Tuple = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['''category'''][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-1_00] * len(UpperCamelCase ), "end_token": [-1_00] * len(UpperCamelCase ), "category": category, }, } _UpperCamelCase : List[str] = out['''context'''].split() _UpperCamelCase : Optional[Any] = splitted_context[answer['''end_token''']] _UpperCamelCase : Optional[int] = len( tokenizer( ''' '''.join(splitted_context[: answer['''start_token''']] ) ,add_special_tokens=UpperCamelCase ,).input_ids ) _UpperCamelCase : Tuple = len( tokenizer(''' '''.join(splitted_context[: answer['''end_token''']] ) ,add_special_tokens=UpperCamelCase ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token _UpperCamelCase : int = len(tokenizer(UpperCamelCase ,add_special_tokens=UpperCamelCase ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 _UpperCamelCase : Optional[Any] = input_ids[answer['''start_token'''] : answer['''end_token'''] + 1] # right & left are inclusive _UpperCamelCase : Any = answer['''start_token'''] _UpperCamelCase : Dict = answer['''end_token'''] if assertion: _UpperCamelCase : Union[str, Any] = tokenizer.decode(UpperCamelCase ) if answer["span"] != new: print('''ISSUE IN TOKENIZATION''' ) print('''OLD:''' ,answer['''span'''] ) print('''NEW:''' ,UpperCamelCase ,end='''\n\n''' ) if len(UpperCamelCase ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } _UpperCamelCase : str = input_ids[:q_len] _UpperCamelCase : int = range(UpperCamelCase ,len(UpperCamelCase ) ,max_length - doc_stride ) _UpperCamelCase : List[Any] = [] _UpperCamelCase : int = [] _UpperCamelCase : Dict = [] _UpperCamelCase : Optional[int] = [] # null, yes, no, long, short for i in doc_start_indices: _UpperCamelCase : Optional[Any] = i + max_length - q_len _UpperCamelCase : int = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: _UpperCamelCase : Optional[int] = start_token - i + q_len _UpperCamelCase : Optional[int] = end_token - i + q_len answers_category.append(answer['''category'''][0] ) # ["short"] -> "short" else: _UpperCamelCase : Optional[Any] = -1_00 _UpperCamelCase : str = -1_00 answers_category.append('''null''' ) _UpperCamelCase : Union[str, Any] = inputs[-1][start_token : end_token + 1] answers_start_token.append(UpperCamelCase ) answers_end_token.append(UpperCamelCase ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('''ISSUE in strided for ID:''' ,example['''id'''] ) print('''New:''' ,tokenizer.decode(UpperCamelCase ) ) print('''Old:''' ,tokenizer.decode(UpperCamelCase ) ,end='''\n\n''' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase=20_48 ,UpperCamelCase=40_96 ,UpperCamelCase=False ) -> Tuple: _UpperCamelCase : Dict = get_strided_contexts_and_ans( UpperCamelCase ,UpperCamelCase ,doc_stride=UpperCamelCase ,max_length=UpperCamelCase ,assertion=UpperCamelCase ,) return example def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Any: with jsonlines.open(UpperCamelCase ,'''a''' ) as writer: for example in tqdm(UpperCamelCase ,total=len(UpperCamelCase ) ,desc='''Saving samples ... ''' ): _UpperCamelCase : Union[str, Any] = example['''labels'''] for ids, start, end, cat in zip( example['''input_ids'''] ,labels['''start_token'''] ,labels['''end_token'''] ,labels['''category'''] ,): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { '''input_ids''': ids, '''start_token''': start, '''end_token''': end, '''category''': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer _UpperCAmelCase : Any = load_dataset("""natural_questions""") _UpperCAmelCase : List[Any] = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") _UpperCAmelCase : Any = data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] _UpperCAmelCase : str = { """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } _UpperCAmelCase : List[str] = data.map(prepare_inputs, fn_kwargs=fn_kwargs) _UpperCAmelCase : int = data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) _UpperCAmelCase : Tuple = """nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
683
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers _UpperCAmelCase : Tuple = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) print("""NCCL version:""", torch.cuda.nccl.version()) except ImportError: print("""Torch version:""", None) try: import deepspeed print("""DeepSpeed version:""", deepspeed.__version__) except ImportError: print("""DeepSpeed version:""", None) try: import tensorflow as tf print("""TensorFlow version:""", tf.__version__) print("""TF GPUs available:""", bool(tf.config.list_physical_devices("""GPU"""))) print("""Number of TF GPUs available:""", len(tf.config.list_physical_devices("""GPU"""))) except ImportError: print("""TensorFlow version:""", None)
683
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { """YituTech/conv-bert-base""": """https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json""", """YituTech/conv-bert-medium-small""": ( """https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json""" ), """YituTech/conv-bert-small""": """https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json""", # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[str] = 'convbert' def __init__( self , _snake_case=30522 , _snake_case=768 , _snake_case=12 , _snake_case=12 , _snake_case=3072 , _snake_case="gelu" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=2 , _snake_case=0.02 , _snake_case=1E-12 , _snake_case=1 , _snake_case=0 , _snake_case=2 , _snake_case=768 , _snake_case=2 , _snake_case=9 , _snake_case=1 , _snake_case=None , **_snake_case , ) -> Union[str, Any]: super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , **_snake_case , ) _UpperCamelCase : int = vocab_size _UpperCamelCase : str = hidden_size _UpperCamelCase : Union[str, Any] = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : Any = intermediate_size _UpperCamelCase : List[Any] = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : List[Any] = attention_probs_dropout_prob _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = type_vocab_size _UpperCamelCase : Optional[Any] = initializer_range _UpperCamelCase : int = layer_norm_eps _UpperCamelCase : Tuple = embedding_size _UpperCamelCase : Optional[Any] = head_ratio _UpperCamelCase : Optional[Any] = conv_kernel_size _UpperCamelCase : Tuple = num_groups _UpperCamelCase : List[str] = classifier_dropout class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCamelCase : str = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCamelCase : Optional[int] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
683
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: return params[f'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase="attention" ) -> List[str]: _UpperCamelCase : Dict = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) _UpperCamelCase : int = k_tmp.reshape(k_tmp.shape[0] ,k_tmp.shape[1] * k_tmp.shape[2] ) _UpperCamelCase : str = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) _UpperCamelCase : Tuple = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] ,o_tmp.shape[2] ) _UpperCamelCase : Any = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) _UpperCamelCase : Optional[int] = q_tmp.reshape(q_tmp.shape[0] ,q_tmp.shape[1] * q_tmp.shape[2] ) _UpperCamelCase : Optional[Any] = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) _UpperCamelCase : List[Any] = v_tmp.reshape(v_tmp.shape[0] ,v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[str]: if split_mlp_wi: _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] _UpperCamelCase : Tuple = params[f'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] _UpperCamelCase : Optional[Any] = (wi_a, wi_a) else: _UpperCamelCase : str = params[f'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: return params[f'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def snake_case__ ( UpperCamelCase ,*, UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ) -> int: _UpperCamelCase : Any = traverse_util.flatten_dict(variables['''target'''] ) _UpperCamelCase : Optional[Any] = {'''/'''.join(UpperCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _UpperCamelCase : str = '''encoder/encoder/mlp/wi_0/kernel''' in old print('''Split MLP:''' ,UpperCamelCase ) _UpperCamelCase : Optional[int] = collections.OrderedDict() # Shared embeddings. _UpperCamelCase : str = old['''token_embedder/embedding'''] # Encoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''attention''' ) _UpperCamelCase : Tuple = layer_norm _UpperCamelCase : int = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : Dict = v.T # Block i, layer 1 (MLP). _UpperCamelCase : Dict = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : int = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,UpperCamelCase ) _UpperCamelCase : Union[str, Any] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Optional[Any] = wi[1].T else: _UpperCamelCase : List[Any] = wi.T _UpperCamelCase : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : Union[str, Any] = tax_relpos_bias_lookup( UpperCamelCase ,UpperCamelCase ,'''encoder''' ).T _UpperCamelCase : List[str] = old['''encoder/encoder_norm/scale'''] if not scalable_attention: _UpperCamelCase : List[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''encoder''' ).T _UpperCamelCase : Optional[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''decoder''' ).T if not is_encoder_only: # Decoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_self_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''self_attention''' ) _UpperCamelCase : int = layer_norm _UpperCamelCase : Union[str, Any] = k.T _UpperCamelCase : Optional[int] = o.T _UpperCamelCase : Dict = q.T _UpperCamelCase : Tuple = v.T # Block i, layer 1 (Cross Attention). _UpperCamelCase : str = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_cross_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Dict = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''encoder_decoder_attention''' ) _UpperCamelCase : Dict = layer_norm _UpperCamelCase : Optional[int] = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : str = v.T # Block i, layer 2 (MLP). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : List[Any] = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,UpperCamelCase ) _UpperCamelCase : List[str] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Union[str, Any] = wi[1].T else: _UpperCamelCase : Dict = wi.T _UpperCamelCase : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : int = tax_relpos_bias_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ).T _UpperCamelCase : Optional[int] = old['''decoder/decoder_norm/scale'''] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _UpperCamelCase : str = old['''decoder/logits_dense/kernel'''].T return new def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Optional[int]: _UpperCamelCase : str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : str = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : int = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''' ) _UpperCamelCase : Any = state_dict['''shared.weight'''] return state_dict def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Any: _UpperCamelCase : List[Any] = checkpoints.load_tax_checkpoint(UpperCamelCase ) _UpperCamelCase : str = convert_tax_to_pytorch( UpperCamelCase ,num_layers=config.num_layers ,is_encoder_only=UpperCamelCase ,scalable_attention=UpperCamelCase ) _UpperCamelCase : Optional[Any] = make_state_dict(UpperCamelCase ,UpperCamelCase ) model.load_state_dict(UpperCamelCase ,strict=UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ,UpperCamelCase = False ,) -> int: _UpperCamelCase : int = MTaConfig.from_json_file(UpperCamelCase ) print(f'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _UpperCamelCase : Optional[int] = UMTaEncoderModel(UpperCamelCase ) else: _UpperCamelCase : Optional[int] = UMTaForConditionalGeneration(UpperCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(UpperCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(UpperCamelCase ) print('''Done''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
683
1
'''simple docstring''' from __future__ import annotations def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: if len(UpperCamelCase ) < k or k < 0: raise ValueError('''Invalid Input''' ) _UpperCamelCase : Tuple = sum(array[:k] ) for i in range(len(UpperCamelCase ) - k ): _UpperCamelCase : List[Any] = current_sum - array[i] + array[i + k] _UpperCamelCase : int = max(UpperCamelCase ,UpperCamelCase ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() _UpperCAmelCase : List[str] = [randint(-1000, 1000) for i in range(100)] _UpperCAmelCase : str = randint(0, 110) print(f"""The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}""")
683
'''simple docstring''' from __future__ import annotations from functools import lru_cache from math import ceil _UpperCAmelCase : int = 100 _UpperCAmelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _UpperCAmelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_00 ) def snake_case__ ( UpperCamelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _UpperCamelCase : set[int] = set() _UpperCamelCase : int _UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def snake_case__ ( UpperCamelCase = 50_00 ) -> int | None: for number_to_partition in range(1 ,UpperCamelCase ): if len(partition(UpperCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
683
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } _UpperCAmelCase : Any = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : Tuple = DistilBertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> int: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : int = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = do_lower_case _UpperCamelCase : Dict = strip_accents _UpperCamelCase : List[Any] = tokenize_chinese_chars _UpperCamelCase : Tuple = normalizer_class(**_snake_case ) _UpperCamelCase : Dict = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[int]: _UpperCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Optional[Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _UpperCAmelCase : Dict = """bart""" _UpperCAmelCase : List[str] = True @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> int: if LOAD_DENSE_INDEX: _UpperCamelCase : Optional[int] = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _UpperCamelCase : Optional[Any] = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _UpperCamelCase : Tuple = qar_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = (None, None) if MODEL_TYPE == "bart": _UpperCamelCase : Any = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _UpperCamelCase : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _UpperCamelCase : Dict = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _UpperCamelCase : Tuple = sas_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = make_qa_sas_model( model_name='''t5-small''' ,from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' ,device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> List[Any]: if LOAD_DENSE_INDEX: _UpperCamelCase : str = faiss.StandardGpuResources() _UpperCamelCase : Optional[int] = datasets.load_dataset(path='''wiki_snippets''' ,name='''wiki40b_en_100_0''' )['''train'''] _UpperCamelCase : List[str] = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(wikiaab_passages.num_rows, 1_28) ,) _UpperCamelCase : Any = faiss.IndexFlatIP(1_28 ) _UpperCamelCase : str = faiss.index_cpu_to_gpu(UpperCamelCase ,1 ,UpperCamelCase ) wikiaab_gpu_index_flat.add(UpperCamelCase ) # TODO fix for larger GPU else: _UpperCamelCase, _UpperCamelCase : Optional[int] = (None, None) _UpperCamelCase : int = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : List[Any] = datasets.load_dataset('''eli5''' ,name='''LFQA_reddit''' ) _UpperCamelCase : Optional[int] = elia['''train_eli5'''] _UpperCamelCase : Any = np.memmap( '''eli5_questions_reps.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(elia_train.num_rows, 1_28) ) _UpperCamelCase : Optional[Any] = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(UpperCamelCase ) return (elia_train, eli5_train_q_index) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_indexes() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_models() _UpperCAmelCase , _UpperCAmelCase : int = load_train_data() def snake_case__ ( UpperCamelCase ,UpperCamelCase=10 ) -> Optional[Any]: _UpperCamelCase : Optional[int] = embed_questions_for_retrieval([question] ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase, _UpperCamelCase : Optional[Any] = eli5_train_q_index.search(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = [elia_train[int(UpperCamelCase )] for i in I[0]] return nn_examples def snake_case__ ( UpperCamelCase ,UpperCamelCase="wiki40b" ,UpperCamelCase="dense" ,UpperCamelCase=10 ) -> Optional[int]: if source == "none": _UpperCamelCase, _UpperCamelCase : Dict = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _UpperCamelCase, _UpperCamelCase : str = query_qa_dense_index( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else: _UpperCamelCase, _UpperCamelCase : str = query_es_index( UpperCamelCase ,UpperCamelCase ,index_name='''english_wiki40b_snippets_100w''' ,n_results=UpperCamelCase ,) _UpperCamelCase : Optional[int] = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _UpperCamelCase : Optional[Any] = '''question: {} context: {}'''.format(UpperCamelCase ,UpperCamelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda UpperCamelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda UpperCamelCase : None), } ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=64 ,UpperCamelCase=2_56 ,UpperCamelCase=False ,UpperCamelCase=2 ,UpperCamelCase=0.95 ,UpperCamelCase=0.8 ) -> Optional[Any]: with torch.no_grad(): _UpperCamelCase : Any = qa_sas_generate( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,num_answers=1 ,num_beams=UpperCamelCase ,min_len=UpperCamelCase ,max_len=UpperCamelCase ,do_sample=UpperCamelCase ,temp=UpperCamelCase ,top_p=UpperCamelCase ,top_k=UpperCamelCase ,max_input_length=10_24 ,device='''cuda:0''' ,)[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar _UpperCAmelCase : str = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _UpperCAmelCase : Tuple = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _UpperCAmelCase : Dict = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _UpperCAmelCase : List[str] = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _UpperCAmelCase : Optional[int] = st.sidebar.checkbox("""Demo options""") if demo_options: _UpperCAmelCase : List[str] = st.sidebar.selectbox( """""", action_list, index=3, ) _UpperCAmelCase : List[Any] = action_list.index(action_st) _UpperCAmelCase : Tuple = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) _UpperCAmelCase : Optional[Any] = show_type == """Show full text of passages""" else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : str = True _UpperCAmelCase : str = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: _UpperCAmelCase : Optional[Any] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) _UpperCAmelCase : Optional[Any] = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: _UpperCAmelCase : Dict = """wiki40b""" _UpperCAmelCase : str = """dense""" _UpperCAmelCase : List[str] = """beam""" _UpperCAmelCase : Dict = 2 _UpperCAmelCase : List[str] = 64 _UpperCAmelCase : List[Any] = 256 _UpperCAmelCase : Tuple = None _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : int = st.sidebar.checkbox("""Generation options""") if generate_options: _UpperCAmelCase : Union[str, Any] = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) _UpperCAmelCase : Dict = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) _UpperCAmelCase : List[Any] = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": _UpperCAmelCase : List[str] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[int] = None # start main text _UpperCAmelCase : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _UpperCAmelCase : int = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": _UpperCAmelCase : Any = st.text_input("""Enter your question here:""", """""") else: _UpperCAmelCase : Tuple = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": _UpperCAmelCase , _UpperCAmelCase : str = make_support(question, source=wiki_source, method="""dense""", n_results=10) _UpperCAmelCase , _UpperCAmelCase : List[Any] = make_support(question, source=wiki_source, method="""sparse""", n_results=10) _UpperCAmelCase : int = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _UpperCAmelCase : int = support_list[:10] _UpperCAmelCase : Tuple = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _UpperCAmelCase , _UpperCAmelCase : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): _UpperCAmelCase : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) _UpperCAmelCase : List[Any] = res[1].strip() if sec_titles == "": _UpperCAmelCase : Optional[int] = """[{}]({})""".format(res[0], wiki_url) else: _UpperCAmelCase : Optional[int] = sec_titles.split(""" & """) _UpperCAmelCase : Tuple = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: _UpperCAmelCase : Dict = find_nearest_training(question) _UpperCAmelCase : List[Any] = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) _UpperCAmelCase : List[Any] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) _UpperCAmelCase : List[Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
683
1
'''simple docstring''' import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class UpperCAmelCase ( a_ , a_ ): """simple docstring""" @register_to_config def __init__( self , _snake_case = 128 , _snake_case = 256 , _snake_case = 2_000.0 , _snake_case = 768 , _snake_case = 12 , _snake_case = 12 , _snake_case = 64 , _snake_case = 2048 , _snake_case = 0.1 , ) -> Tuple: super().__init__() _UpperCamelCase : Optional[Any] = nn.Sequential( nn.Linear(_snake_case , d_model * 4 , bias=_snake_case ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_snake_case ) , nn.SiLU() , ) _UpperCamelCase : List[Any] = nn.Embedding(_snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = False _UpperCamelCase : Optional[int] = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) _UpperCamelCase : Optional[int] = nn.Dropout(p=_snake_case ) _UpperCamelCase : int = nn.ModuleList() for lyr_num in range(_snake_case ): # FiLM conditional T5 decoder _UpperCamelCase : Tuple = DecoderLayer(d_model=_snake_case , d_kv=_snake_case , num_heads=_snake_case , d_ff=_snake_case , dropout_rate=_snake_case ) self.decoders.append(_snake_case ) _UpperCamelCase : List[str] = TaLayerNorm(_snake_case ) _UpperCamelCase : List[str] = nn.Dropout(p=_snake_case ) _UpperCamelCase : Any = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Any = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> str: _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[int] = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. _UpperCamelCase : Dict = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) _UpperCamelCase : Any = self.conditioning_emb(_snake_case ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) _UpperCamelCase : Optional[Any] = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. _UpperCamelCase : Union[str, Any] = torch.broadcast_to( torch.arange(_snake_case , device=decoder_input_tokens.device ) , (batch, seq_length) , ) _UpperCamelCase : Union[str, Any] = self.position_encoding(_snake_case ) _UpperCamelCase : List[Any] = self.continuous_inputs_projection(_snake_case ) inputs += position_encodings _UpperCamelCase : Tuple = self.dropout(_snake_case ) # decoder: No padding present. _UpperCamelCase : Optional[int] = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. _UpperCamelCase : Optional[int] = [(x, self.encoder_decoder_mask(_snake_case , _snake_case )) for x, y in encodings_and_masks] # cross attend style: concat encodings _UpperCamelCase : Optional[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) _UpperCamelCase : str = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: _UpperCamelCase : Optional[int] = lyr( _snake_case , conditioning_emb=_snake_case , encoder_hidden_states=_snake_case , encoder_attention_mask=_snake_case , )[0] _UpperCamelCase : List[str] = self.decoder_norm(_snake_case ) _UpperCamelCase : int = self.post_dropout(_snake_case ) _UpperCamelCase : Dict = self.spec_out(_snake_case ) return spec_out class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=1E-6 ) -> str: super().__init__() _UpperCamelCase : int = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_snake_case , d_kv=_snake_case , num_heads=_snake_case , dropout_rate=_snake_case ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_snake_case , d_kv=_snake_case , num_heads=_snake_case , dropout_rate=_snake_case , layer_norm_epsilon=_snake_case , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_snake_case , d_ff=_snake_case , dropout_rate=_snake_case , layer_norm_epsilon=_snake_case ) ) def _lowercase ( self , _snake_case , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case=None , ) -> Optional[Any]: _UpperCamelCase : int = self.layer[0]( _snake_case , conditioning_emb=_snake_case , attention_mask=_snake_case , ) if encoder_hidden_states is not None: _UpperCamelCase : Optional[int] = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) _UpperCamelCase : int = self.layer[1]( _snake_case , key_value_states=_snake_case , attention_mask=_snake_case , ) # Apply Film Conditional Feed Forward layer _UpperCamelCase : Dict = self.layer[-1](_snake_case , _snake_case ) return (hidden_states,) class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> List[Any]: super().__init__() _UpperCamelCase : Any = TaLayerNorm(_snake_case ) _UpperCamelCase : int = TaFiLMLayer(in_features=d_model * 4 , out_features=_snake_case ) _UpperCamelCase : Dict = Attention(query_dim=_snake_case , heads=_snake_case , dim_head=_snake_case , out_bias=_snake_case , scale_qk=_snake_case ) _UpperCamelCase : List[Any] = nn.Dropout(_snake_case ) def _lowercase ( self , _snake_case , _snake_case=None , _snake_case=None , ) -> List[str]: # pre_self_attention_layer_norm _UpperCamelCase : Optional[Any] = self.layer_norm(_snake_case ) if conditioning_emb is not None: _UpperCamelCase : Optional[int] = self.FiLMLayer(_snake_case , _snake_case ) # Self-attention block _UpperCamelCase : str = self.attention(_snake_case ) _UpperCamelCase : List[Any] = hidden_states + self.dropout(_snake_case ) return hidden_states class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) -> List[str]: super().__init__() _UpperCamelCase : Tuple = Attention(query_dim=_snake_case , heads=_snake_case , dim_head=_snake_case , out_bias=_snake_case , scale_qk=_snake_case ) _UpperCamelCase : str = TaLayerNorm(_snake_case , eps=_snake_case ) _UpperCamelCase : Tuple = nn.Dropout(_snake_case ) def _lowercase ( self , _snake_case , _snake_case=None , _snake_case=None , ) -> List[Any]: _UpperCamelCase : Tuple = self.layer_norm(_snake_case ) _UpperCamelCase : Tuple = self.attention( _snake_case , encoder_hidden_states=_snake_case , attention_mask=attention_mask.squeeze(1 ) , ) _UpperCamelCase : Union[str, Any] = hidden_states + self.dropout(_snake_case ) return layer_output class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> Any: super().__init__() _UpperCamelCase : int = TaDenseGatedActDense(d_model=_snake_case , d_ff=_snake_case , dropout_rate=_snake_case ) _UpperCamelCase : Union[str, Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_snake_case ) _UpperCamelCase : Any = TaLayerNorm(_snake_case , eps=_snake_case ) _UpperCamelCase : Dict = nn.Dropout(_snake_case ) def _lowercase ( self , _snake_case , _snake_case=None ) -> Any: _UpperCamelCase : List[Any] = self.layer_norm(_snake_case ) if conditioning_emb is not None: _UpperCamelCase : List[str] = self.film(_snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = self.DenseReluDense(_snake_case ) _UpperCamelCase : Any = hidden_states + self.dropout(_snake_case ) return hidden_states class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case ) -> Dict: super().__init__() _UpperCamelCase : Tuple = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) _UpperCamelCase : int = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) _UpperCamelCase : Dict = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) _UpperCamelCase : int = nn.Dropout(_snake_case ) _UpperCamelCase : Optional[int] = NewGELUActivation() def _lowercase ( self , _snake_case ) -> Dict: _UpperCamelCase : int = self.act(self.wi_a(_snake_case ) ) _UpperCamelCase : List[Any] = self.wi_a(_snake_case ) _UpperCamelCase : Any = hidden_gelu * hidden_linear _UpperCamelCase : Optional[Any] = self.dropout(_snake_case ) _UpperCamelCase : Tuple = self.wo(_snake_case ) return hidden_states class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case=1E-6 ) -> Optional[Any]: super().__init__() _UpperCamelCase : Dict = nn.Parameter(torch.ones(_snake_case ) ) _UpperCamelCase : int = eps def _lowercase ( self , _snake_case ) -> int: # T5 uses a layer_norm which only scales and doesn't shift, which is also known as Root Mean # Square Layer Normalization https://arxiv.org/abs/1910.07467 thus variance is calculated # w/o mean and there is no bias. Additionally we want to make sure that the accumulation for # half-precision inputs is done in fp32 _UpperCamelCase : List[str] = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_snake_case ) _UpperCamelCase : Any = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: _UpperCamelCase : Any = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class UpperCAmelCase ( nn.Module ): """simple docstring""" def _lowercase ( self , _snake_case ) -> torch.Tensor: return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.044_715 * torch.pow(_snake_case , 3.0 )) )) class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case ) -> str: super().__init__() _UpperCamelCase : Tuple = nn.Linear(_snake_case , out_features * 2 , bias=_snake_case ) def _lowercase ( self , _snake_case , _snake_case ) -> Optional[int]: _UpperCamelCase : Tuple = self.scale_bias(_snake_case ) _UpperCamelCase, _UpperCamelCase : Any = torch.chunk(_snake_case , 2 , -1 ) _UpperCamelCase : str = x * (1 + scale) + shift return x
683
'''simple docstring''' from collections.abc import Iterable from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> Optional[int]: _UpperCamelCase : int = value _UpperCamelCase : Node | None = None # Added in order to delete a node easier _UpperCamelCase : Node | None = None _UpperCamelCase : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> List[Any]: _UpperCamelCase : str = root def __str__( self ) -> str: return str(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if new_children is not None: # reset its kids _UpperCamelCase : Union[str, Any] = node.parent if node.parent is not None: # reset its parent if self.is_right(_snake_case ): # If it is the right children _UpperCamelCase : str = new_children else: _UpperCamelCase : Any = new_children else: _UpperCamelCase : Any = new_children def _lowercase ( self , _snake_case ) -> bool: if node.parent and node.parent.right: return node == node.parent.right return False def _lowercase ( self ) -> bool: return self.root is None def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : List[Any] = Node(_snake_case ) # create a new Node if self.empty(): # if Tree is empty _UpperCamelCase : Optional[Any] = new_node # set its root else: # Tree is not empty _UpperCamelCase : int = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: _UpperCamelCase : Union[str, Any] = new_node # We insert the new node in a leaf break else: _UpperCamelCase : Union[str, Any] = parent_node.left else: if parent_node.right is None: _UpperCamelCase : Any = new_node break else: _UpperCamelCase : str = parent_node.right _UpperCamelCase : Any = parent_node def _lowercase ( self , *_snake_case ) -> None: for value in values: self.__insert(_snake_case ) def _lowercase ( self , _snake_case ) -> Node | None: if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: _UpperCamelCase : List[str] = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: _UpperCamelCase : Optional[Any] = node.left if value < node.value else node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: if self.root is None: return None _UpperCamelCase : Dict = self.root if not self.empty(): while node.right is not None: _UpperCamelCase : Tuple = node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: _UpperCamelCase : Optional[Any] = self.root if self.root is None: return None if not self.empty(): _UpperCamelCase : Optional[int] = self.root while node.left is not None: _UpperCamelCase : List[str] = node.left return node def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : str = self.search(_snake_case ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_snake_case , _snake_case ) elif node.left is None: # Has only right children self.__reassign_nodes(_snake_case , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_snake_case , node.left ) else: _UpperCamelCase : List[str] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore _UpperCamelCase : int = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _lowercase ( self , _snake_case ) -> Iterable: if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _lowercase ( self , _snake_case=None ) -> Any: if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if node: self.inorder(_snake_case , node.left ) arr.append(node.value ) self.inorder(_snake_case , node.right ) def _lowercase ( self , _snake_case , _snake_case ) -> int: _UpperCamelCase : list[int] = [] self.inorder(_snake_case , _snake_case ) # append all values to list using inorder traversal return arr[k - 1] def snake_case__ ( UpperCamelCase ) -> list[Node]: _UpperCamelCase : int = [] if curr_node is not None: _UpperCamelCase : Any = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def snake_case__ ( ) -> None: _UpperCamelCase : Any = (8, 3, 6, 1, 10, 14, 13, 4, 7) _UpperCamelCase : Tuple = BinarySearchTree() for i in testlist: t.insert(UpperCamelCase ) # Prints all the elements of the list in order traversal print(UpperCamelCase ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' ,t.get_max().value ) # type: ignore print('''Min Value: ''' ,t.get_min().value ) # type: ignore for i in testlist: t.remove(UpperCamelCase ) print(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
683
1
'''simple docstring''' import math def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> float: return math.pow(UpperCamelCase ,2 ) - a def snake_case__ ( UpperCamelCase ) -> float: return 2 * x def snake_case__ ( UpperCamelCase ) -> float: _UpperCamelCase : Dict = 2.0 while start <= a: _UpperCamelCase : Optional[int] = math.pow(UpperCamelCase ,2 ) return start def snake_case__ ( UpperCamelCase ,UpperCamelCase = 99_99 ,UpperCamelCase = 0.00000000000001 ) -> float: if a < 0: raise ValueError('''math domain error''' ) _UpperCamelCase : Any = get_initial_point(UpperCamelCase ) for _ in range(UpperCamelCase ): _UpperCamelCase : str = value _UpperCamelCase : List[Any] = value - fx(UpperCamelCase ,UpperCamelCase ) / fx_derivative(UpperCamelCase ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
683
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off _UpperCAmelCase : Dict = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 10563, 10786, 11420, 11709, 11907, 13163, 13697, 13700, 14808, 15306, 16410, 16791, 17992, 19203, 19510, 20724, 22305, 22935, 27007, 30109, 30420, 33409, 34949, 40283, 40493, 40549, 47282, 49146, 50257, 50359, 50360, 50361 ] _UpperCAmelCase : int = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362 ] class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = 'whisper' A__ : Tuple = ['past_key_values'] A__ : Optional[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , _snake_case=51865 , _snake_case=80 , _snake_case=6 , _snake_case=4 , _snake_case=6 , _snake_case=4 , _snake_case=1536 , _snake_case=1536 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=50257 , _snake_case=True , _snake_case=True , _snake_case="gelu" , _snake_case=256 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.02 , _snake_case=False , _snake_case=1500 , _snake_case=448 , _snake_case=50256 , _snake_case=50256 , _snake_case=50256 , _snake_case=None , _snake_case=[220, 50256] , _snake_case=False , _snake_case=256 , _snake_case=False , _snake_case=0.05 , _snake_case=10 , _snake_case=2 , _snake_case=0.0 , _snake_case=10 , _snake_case=0 , _snake_case=7 , **_snake_case , ) -> Any: _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Union[str, Any] = num_mel_bins _UpperCamelCase : List[str] = d_model _UpperCamelCase : str = encoder_layers _UpperCamelCase : Optional[int] = encoder_attention_heads _UpperCamelCase : str = decoder_layers _UpperCamelCase : Tuple = decoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : Optional[int] = encoder_ffn_dim _UpperCamelCase : Any = dropout _UpperCamelCase : Optional[Any] = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : int = activation_function _UpperCamelCase : List[Any] = init_std _UpperCamelCase : Optional[int] = encoder_layerdrop _UpperCamelCase : str = decoder_layerdrop _UpperCamelCase : List[str] = use_cache _UpperCamelCase : Optional[Any] = encoder_layers _UpperCamelCase : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : List[str] = max_source_positions _UpperCamelCase : Optional[Any] = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. _UpperCamelCase : str = classifier_proj_size _UpperCamelCase : List[str] = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCamelCase : int = apply_spec_augment _UpperCamelCase : str = mask_time_prob _UpperCamelCase : int = mask_time_length _UpperCamelCase : List[Any] = mask_time_min_masks _UpperCamelCase : List[str] = mask_feature_prob _UpperCamelCase : Optional[int] = mask_feature_length _UpperCamelCase : Union[str, Any] = mask_feature_min_masks _UpperCamelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , is_encoder_decoder=_snake_case , decoder_start_token_id=_snake_case , suppress_tokens=_snake_case , begin_suppress_tokens=_snake_case , **_snake_case , ) class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: _UpperCamelCase : Dict = OrderedDict( [ ('''input_features''', {0: '''batch''', 1: '''feature_size''', 2: '''encoder_sequence'''}), ] ) if self.use_past: _UpperCamelCase : Tuple = {0: '''batch'''} else: _UpperCamelCase : Dict = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_snake_case , direction='''inputs''' ) return common_inputs def _lowercase ( self , _snake_case , _snake_case = -1 , _snake_case = -1 , _snake_case = False , _snake_case = None , _snake_case = 22050 , _snake_case = 5.0 , _snake_case = 220 , ) -> Mapping[str, Any]: _UpperCamelCase : Optional[int] = OrderedDict() _UpperCamelCase : Tuple = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_snake_case , framework=_snake_case , sampling_rate=_snake_case , time_duration=_snake_case , frequency=_snake_case , ) _UpperCamelCase : int = encoder_inputs['''input_features'''].shape[2] _UpperCamelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length _UpperCamelCase : str = super().generate_dummy_inputs( preprocessor.tokenizer , _snake_case , _snake_case , _snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = encoder_inputs.pop('''input_features''' ) _UpperCamelCase : Dict = decoder_inputs.pop('''decoder_input_ids''' ) if "past_key_values" in decoder_inputs: _UpperCamelCase : List[str] = decoder_inputs.pop('''past_key_values''' ) return dummy_inputs @property def _lowercase ( self ) -> float: return 1E-3
683
1
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class UpperCAmelCase ( a_ ): """simple docstring""" def _lowercase ( self , _snake_case ) -> float: return 0.0 def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> tuple[int | float, int | float]: _UpperCamelCase : Optional[int] = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _UpperCamelCase : Optional[int] = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> None: _UpperCamelCase : int = 5_12 _UpperCamelCase : Dict = [1] + [0] * (size - 1) _UpperCamelCase : Union[str, Any] = [filter_type.process(UpperCamelCase ) for item in inputs] _UpperCamelCase : int = [0] * (samplerate - size) # zero-padding outputs += filler _UpperCamelCase : Any = np.abs(np.fft.fft(UpperCamelCase ) ) _UpperCamelCase : Any = 20 * np.logaa(UpperCamelCase ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 ,samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) # Display within reasonable bounds _UpperCamelCase : str = get_bounds(UpperCamelCase ,UpperCamelCase ) plt.ylim(max([-80, bounds[0]] ) ,min([80, bounds[1]] ) ) plt.ylabel('''Gain (dB)''' ) plt.plot(UpperCamelCase ) plt.show() def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> None: _UpperCamelCase : Dict = 5_12 _UpperCamelCase : Union[str, Any] = [1] + [0] * (size - 1) _UpperCamelCase : List[str] = [filter_type.process(UpperCamelCase ) for item in inputs] _UpperCamelCase : Dict = [0] * (samplerate - size) # zero-padding outputs += filler _UpperCamelCase : int = np.angle(np.fft.fft(UpperCamelCase ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 ,samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) plt.ylim(-2 * pi ,2 * pi ) plt.ylabel('''Phase shift (Radians)''' ) plt.plot(np.unwrap(UpperCamelCase ,-2 * pi ) ) plt.show()
683
'''simple docstring''' import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser( description=( """Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""roberta""", choices=["""roberta""", """gpt2"""]) parser.add_argument("""--model_name""", default="""roberta-large""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_roberta_048131723.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") _UpperCAmelCase : int = parser.parse_args() if args.model_type == "roberta": _UpperCAmelCase : Union[str, Any] = RobertaForMaskedLM.from_pretrained(args.model_name) _UpperCAmelCase : int = """roberta""" elif args.model_type == "gpt2": _UpperCAmelCase : Optional[int] = GPTaLMHeadModel.from_pretrained(args.model_name) _UpperCAmelCase : Optional[int] = """transformer""" _UpperCAmelCase : Tuple = model.state_dict() _UpperCAmelCase : int = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: _UpperCAmelCase : Optional[Any] = state_dict[f"""{prefix}.{param_name}"""] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: _UpperCAmelCase : Tuple = f"""{prefix}.embeddings.{w}.weight""" _UpperCAmelCase : Optional[Any] = state_dict[param_name] for w in ["weight", "bias"]: _UpperCAmelCase : Union[str, Any] = f"""{prefix}.embeddings.LayerNorm.{w}""" _UpperCAmelCase : str = state_dict[param_name] # Transformer Blocks # _UpperCAmelCase : Dict = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: _UpperCAmelCase : str = state_dict[ f"""{prefix}.h.{teacher_idx}.{layer}.{w}""" ] _UpperCAmelCase : Any = state_dict[f"""{prefix}.h.{teacher_idx}.attn.bias"""] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: _UpperCAmelCase : Optional[Any] = state_dict[ f"""{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}""" ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: _UpperCAmelCase : Dict = state_dict[f"""{layer}"""] if args.vocab_transform: for w in ["weight", "bias"]: _UpperCAmelCase : int = state_dict[f"""lm_head.dense.{w}"""] _UpperCAmelCase : int = state_dict[f"""lm_head.layer_norm.{w}"""] elif args.model_type == "gpt2": for w in ["weight", "bias"]: _UpperCAmelCase : List[str] = state_dict[f"""{prefix}.ln_f.{w}"""] _UpperCAmelCase : Any = state_dict["""lm_head.weight"""] print(f"""N layers selected for distillation: {std_idx}""") print(f"""Number of params transferred for distillation: {len(compressed_sd.keys())}""") print(f"""Save transferred checkpoint to {args.dump_checkpoint}.""") torch.save(compressed_sd, args.dump_checkpoint)
683
1
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=a_ ) class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = field(default='summarization' , metadata={'include_in_asdict_even_if_is_default': True} ) A__ : ClassVar[Features] = Features({'text': Value('string' )} ) A__ : ClassVar[Features] = Features({'summary': Value('string' )} ) A__ : str = "text" A__ : str = "summary" @property def _lowercase ( self ) -> Dict[str, str]: return {self.text_column: "text", self.summary_column: "summary"}
683
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = None _UpperCamelCase : int = 20 _UpperCamelCase : Any = self._get_uniform_logits(batch_size=2 , length=_snake_case ) # tweak scores to not be uniform anymore _UpperCamelCase : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch _UpperCamelCase : Dict = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax _UpperCamelCase : Any = jax.nn.softmax(_snake_case , axis=-1 ) _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : List[str] = FlaxTemperatureLogitsWarper(temperature=1.3 ) _UpperCamelCase : List[str] = jax.nn.softmax(temp_dist_warper_sharper(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) _UpperCamelCase : str = jax.nn.softmax(temp_dist_warper_smoother(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _lowercase ( self ) -> Any: _UpperCamelCase : List[Any] = None _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : Any = 2 # create ramp distribution _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() _UpperCamelCase : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case _UpperCamelCase : Optional[int] = 5 _UpperCamelCase : str = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) _UpperCamelCase : Union[str, Any] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, length) ).copy() _UpperCamelCase : Optional[Any] = top_k_warp_safety_check(_snake_case , _snake_case , cur_len=_snake_case ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = None _UpperCamelCase : Any = 10 _UpperCamelCase : List[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) _UpperCamelCase : Tuple = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) _UpperCamelCase : List[str] = FlaxTopPLogitsWarper(0.8 ) _UpperCamelCase : Dict = np.exp(top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 _UpperCamelCase : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # check edge cases with negative and extreme logits _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme _UpperCamelCase : Tuple = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept _UpperCamelCase : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) _UpperCamelCase : Dict = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = 20 _UpperCamelCase : Optional[int] = 4 _UpperCamelCase : int = 0 _UpperCamelCase : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) # check that min length is applied at length 5 _UpperCamelCase : Any = ids_tensor((batch_size, 20) , vocab_size=20 ) _UpperCamelCase : int = 5 _UpperCamelCase : List[Any] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 _UpperCamelCase : Optional[int] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = 15 _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = 20 _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) # check that all scores are -inf except the bos_token_id score _UpperCamelCase : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) _UpperCamelCase : Optional[int] = 1 _UpperCamelCase : str = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : str = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 _UpperCamelCase : List[str] = 3 _UpperCamelCase : Tuple = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 20 _UpperCamelCase : Tuple = 4 _UpperCamelCase : Any = 0 _UpperCamelCase : str = 5 _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) # check that all scores are -inf except the eos_token_id when max_length is reached _UpperCamelCase : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) _UpperCamelCase : Dict = 4 _UpperCamelCase : Dict = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : int = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Any = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 4 _UpperCamelCase : Optional[Any] = 10 _UpperCamelCase : Dict = 15 _UpperCamelCase : Union[str, Any] = 2 _UpperCamelCase : Optional[Any] = 1 _UpperCamelCase : List[Any] = 15 # dummy input_ids and scores _UpperCamelCase : Optional[int] = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Any = input_ids.copy() _UpperCamelCase : int = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Tuple = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Optional[int] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : int = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : List[str] = 10 # no processor list _UpperCamelCase : Dict = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) # with processor list _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : Optional[Any] = processor(_snake_case , _snake_case , cur_len=_snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = 4 _UpperCamelCase : int = 10 _UpperCamelCase : List[Any] = 15 _UpperCamelCase : Dict = 2 _UpperCamelCase : Tuple = 1 _UpperCamelCase : Optional[int] = 15 # dummy input_ids and scores _UpperCamelCase : Tuple = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Optional[Any] = input_ids.copy() _UpperCamelCase : List[str] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[int] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : int = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : Union[str, Any] = 10 # no processor list def run_no_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : List[Any] = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) return scores # with processor list def run_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : List[str] = processor(_snake_case , _snake_case , cur_len=_snake_case ) return scores _UpperCamelCase : Dict = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jitted_run_no_processor_list(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = jitted_run_processor_list(_snake_case , _snake_case , _snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
683
1
'''simple docstring''' _UpperCAmelCase : Dict = { 0: """0""", 1: """1""", 2: """2""", 3: """3""", 4: """4""", 5: """5""", 6: """6""", 7: """7""", 8: """8""", 9: """9""", 10: """a""", 11: """b""", 12: """c""", 13: """d""", 14: """e""", 15: """f""", } def snake_case__ ( UpperCamelCase ) -> str: assert type(UpperCamelCase ) in (int, float) and decimal == int(UpperCamelCase ) _UpperCamelCase : Dict = int(UpperCamelCase ) _UpperCamelCase : str = '''''' _UpperCamelCase : Union[str, Any] = False if decimal < 0: _UpperCamelCase : Tuple = True decimal *= -1 while decimal > 0: _UpperCamelCase, _UpperCamelCase : Union[str, Any] = divmod(UpperCamelCase ,16 ) _UpperCamelCase : int = values[remainder] + hexadecimal _UpperCamelCase : Union[str, Any] = '''0x''' + hexadecimal if negative: _UpperCamelCase : Tuple = '''-''' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Optional[int] = pytest.mark.integration @pytest.mark.parametrize('''path''' ,['''paws''', '''csv'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: inspect_dataset(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' ,['''accuracy'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: inspect_metric(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[str] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[str]: with pytest.raises(UpperCamelCase ): get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) @pytest.mark.parametrize( '''path, expected''' ,[ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : int = get_dataset_config_names(UpperCamelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' ,[ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Dict = get_dataset_infos(UpperCamelCase ) assert list(infos.keys() ) == expected_configs _UpperCamelCase : Dict = expected_configs[0] assert expected_config in infos _UpperCamelCase : Any = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : List[Any] = get_dataset_infos(UpperCamelCase ) assert expected_config in infos _UpperCamelCase : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: with pytest.raises(UpperCamelCase ): get_dataset_split_names(UpperCamelCase ,config_name=UpperCamelCase )
683
1
'''simple docstring''' from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig _UpperCAmelCase : Dict = logging.get_logger(__name__) # General docstring _UpperCAmelCase : Tuple = """RegNetConfig""" # Base docstring _UpperCAmelCase : Union[str, Any] = """facebook/regnet-y-040""" _UpperCAmelCase : List[str] = [1, 1088, 7, 7] # Image classification docstring _UpperCAmelCase : List[str] = """facebook/regnet-y-040""" _UpperCAmelCase : Union[str, Any] = """tabby, tabby cat""" _UpperCAmelCase : Optional[int] = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case = 3 , _snake_case = 1 , _snake_case = 1 , _snake_case = "relu" , ) -> Union[str, Any]: super().__init__() _UpperCamelCase : Union[str, Any] = nn.Convad( _snake_case , _snake_case , kernel_size=_snake_case , stride=_snake_case , padding=kernel_size // 2 , groups=_snake_case , bias=_snake_case , ) _UpperCamelCase : List[str] = nn.BatchNormad(_snake_case ) _UpperCamelCase : Union[str, Any] = ACTaFN[activation] if activation is not None else nn.Identity() def _lowercase ( self , _snake_case ) -> Union[str, Any]: _UpperCamelCase : List[str] = self.convolution(_snake_case ) _UpperCamelCase : int = self.normalization(_snake_case ) _UpperCamelCase : int = self.activation(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case ) -> Dict: super().__init__() _UpperCamelCase : int = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) _UpperCamelCase : Any = config.num_channels def _lowercase ( self , _snake_case ) -> Any: _UpperCamelCase : int = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) _UpperCamelCase : Optional[int] = self.embedder(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case = 2 ) -> Optional[int]: super().__init__() _UpperCamelCase : Dict = nn.Convad(_snake_case , _snake_case , kernel_size=1 , stride=_snake_case , bias=_snake_case ) _UpperCamelCase : List[Any] = nn.BatchNormad(_snake_case ) def _lowercase ( self , _snake_case ) -> Tensor: _UpperCamelCase : Optional[int] = self.convolution(_snake_case ) _UpperCamelCase : Optional[int] = self.normalization(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case ) -> Dict: super().__init__() _UpperCamelCase : Union[str, Any] = nn.AdaptiveAvgPoolad((1, 1) ) _UpperCamelCase : Optional[Any] = nn.Sequential( nn.Convad(_snake_case , _snake_case , kernel_size=1 ) , nn.ReLU() , nn.Convad(_snake_case , _snake_case , kernel_size=1 ) , nn.Sigmoid() , ) def _lowercase ( self , _snake_case ) -> Tuple: # b c h w -> b c 1 1 _UpperCamelCase : Optional[Any] = self.pooler(_snake_case ) _UpperCamelCase : Optional[int] = self.attention(_snake_case ) _UpperCamelCase : Dict = hidden_state * attention return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case = 1 ) -> Dict: super().__init__() _UpperCamelCase : int = in_channels != out_channels or stride != 1 _UpperCamelCase : Dict = max(1 , out_channels // config.groups_width ) _UpperCamelCase : List[Any] = ( RegNetShortCut(_snake_case , _snake_case , stride=_snake_case ) if should_apply_shortcut else nn.Identity() ) _UpperCamelCase : str = nn.Sequential( RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , stride=_snake_case , groups=_snake_case , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=_snake_case ) , ) _UpperCamelCase : Any = ACTaFN[config.hidden_act] def _lowercase ( self , _snake_case ) -> Any: _UpperCamelCase : Any = hidden_state _UpperCamelCase : Optional[int] = self.layer(_snake_case ) _UpperCamelCase : Any = self.shortcut(_snake_case ) hidden_state += residual _UpperCamelCase : List[str] = self.activation(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case = 1 ) -> Tuple: super().__init__() _UpperCamelCase : List[Any] = in_channels != out_channels or stride != 1 _UpperCamelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCamelCase : Tuple = ( RegNetShortCut(_snake_case , _snake_case , stride=_snake_case ) if should_apply_shortcut else nn.Identity() ) _UpperCamelCase : Union[str, Any] = nn.Sequential( RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , stride=_snake_case , groups=_snake_case , activation=config.hidden_act ) , RegNetSELayer(_snake_case , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=_snake_case ) , ) _UpperCamelCase : int = ACTaFN[config.hidden_act] def _lowercase ( self , _snake_case ) -> int: _UpperCamelCase : List[Any] = hidden_state _UpperCamelCase : Dict = self.layer(_snake_case ) _UpperCamelCase : Tuple = self.shortcut(_snake_case ) hidden_state += residual _UpperCamelCase : Union[str, Any] = self.activation(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case = 2 , _snake_case = 2 , ) -> Any: super().__init__() _UpperCamelCase : Optional[Any] = RegNetXLayer if config.layer_type == '''x''' else RegNetYLayer _UpperCamelCase : List[Any] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( _snake_case , _snake_case , _snake_case , stride=_snake_case , ) , *[layer(_snake_case , _snake_case , _snake_case ) for _ in range(depth - 1 )] , ) def _lowercase ( self , _snake_case ) -> List[str]: _UpperCamelCase : Tuple = self.layers(_snake_case ) return hidden_state class UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _snake_case ) -> str: super().__init__() _UpperCamelCase : List[str] = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( _snake_case , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) _UpperCamelCase : Tuple = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_snake_case , config.depths[1:] ): self.stages.append(RegNetStage(_snake_case , _snake_case , _snake_case , depth=_snake_case ) ) def _lowercase ( self , _snake_case , _snake_case = False , _snake_case = True ) -> BaseModelOutputWithNoAttention: _UpperCamelCase : int = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCamelCase : List[str] = hidden_states + (hidden_state,) _UpperCamelCase : Any = stage_module(_snake_case ) if output_hidden_states: _UpperCamelCase : Union[str, Any] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=_snake_case , hidden_states=_snake_case ) class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Optional[Any] = RegNetConfig A__ : Dict = 'regnet' A__ : str = 'pixel_values' A__ : Dict = True def _lowercase ( self , _snake_case ) -> Tuple: if isinstance(_snake_case , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''' ) elif isinstance(_snake_case , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _lowercase ( self , _snake_case , _snake_case=False ) -> Optional[int]: if isinstance(_snake_case , _snake_case ): _UpperCamelCase : List[str] = value _UpperCAmelCase : List[Any] = R""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ _UpperCAmelCase : Any = R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( 'The bare RegNet model outputting raw features without any specific head on top.' , a_ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class UpperCAmelCase ( a_ ): """simple docstring""" def __init__( self , _snake_case ) -> List[str]: super().__init__(_snake_case ) _UpperCamelCase : Optional[int] = config _UpperCamelCase : Any = RegNetEmbeddings(_snake_case ) _UpperCamelCase : str = RegNetEncoder(_snake_case ) _UpperCamelCase : List[Any] = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_snake_case ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_snake_case , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _lowercase ( self , _snake_case , _snake_case = None , _snake_case = None ) -> BaseModelOutputWithPoolingAndNoAttention: _UpperCamelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCamelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _UpperCamelCase : Optional[int] = self.embedder(_snake_case ) _UpperCamelCase : Union[str, Any] = self.encoder( _snake_case , output_hidden_states=_snake_case , return_dict=_snake_case ) _UpperCamelCase : Dict = encoder_outputs[0] _UpperCamelCase : int = self.pooler(_snake_case ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_snake_case , pooler_output=_snake_case , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ' , a_ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class UpperCAmelCase ( a_ ): """simple docstring""" def __init__( self , _snake_case ) -> Optional[Any]: super().__init__(_snake_case ) _UpperCamelCase : List[str] = config.num_labels _UpperCamelCase : List[str] = RegNetModel(_snake_case ) # classification head _UpperCamelCase : Optional[Any] = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_snake_case ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_snake_case , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _lowercase ( self , _snake_case = None , _snake_case = None , _snake_case = None , _snake_case = None , ) -> ImageClassifierOutputWithNoAttention: _UpperCamelCase : Optional[int] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCamelCase : Optional[int] = self.regnet(_snake_case , output_hidden_states=_snake_case , return_dict=_snake_case ) _UpperCamelCase : Dict = outputs.pooler_output if return_dict else outputs[1] _UpperCamelCase : Optional[Any] = self.classifier(_snake_case ) _UpperCamelCase : str = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _UpperCamelCase : Union[str, Any] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _UpperCamelCase : Any = '''single_label_classification''' else: _UpperCamelCase : int = '''multi_label_classification''' if self.config.problem_type == "regression": _UpperCamelCase : int = MSELoss() if self.num_labels == 1: _UpperCamelCase : str = loss_fct(logits.squeeze() , labels.squeeze() ) else: _UpperCamelCase : Union[str, Any] = loss_fct(_snake_case , _snake_case ) elif self.config.problem_type == "single_label_classification": _UpperCamelCase : int = CrossEntropyLoss() _UpperCamelCase : int = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _UpperCamelCase : Optional[int] = BCEWithLogitsLoss() _UpperCamelCase : Optional[Any] = loss_fct(_snake_case , _snake_case ) if not return_dict: _UpperCamelCase : Union[str, Any] = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_snake_case , logits=_snake_case , hidden_states=outputs.hidden_states )
683
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ) -> Dict: torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return model @property def _lowercase ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , cross_attention_dim=10 , ) return model @property def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') , up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') , ) _UpperCamelCase : int = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return vqvae, unet @slow def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Tuple = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _UpperCamelCase : int = DDPMScheduler() _UpperCamelCase : Optional[int] = AudioDiffusionPipeline(vqvae=_snake_case , unet=self.dummy_unet , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case , steps=4 ) _UpperCamelCase : Union[str, Any] = output.audios[0] _UpperCamelCase : Union[str, Any] = output.images[0] _UpperCamelCase : str = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : int = pipe(generator=_snake_case , steps=4 , return_dict=_snake_case ) _UpperCamelCase : int = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _UpperCamelCase : List[str] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : List[str] = np.frombuffer(image_from_tuple.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : int = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : str = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _UpperCamelCase : Dict = DDIMScheduler() _UpperCamelCase : str = self.dummy_vqvae_and_unet _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : Optional[Any] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _UpperCamelCase : Optional[Any] = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Tuple = pipe(raw_audio=_snake_case , generator=_snake_case , start_step=5 , steps=10 ) _UpperCamelCase : List[str] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _UpperCamelCase : Any = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Tuple = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : Any = self.dummy_unet_condition _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_snake_case , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : Union[str, Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : int = torch.rand((1, 1, 10) ) _UpperCamelCase : Optional[Any] = pipe(generator=_snake_case , encoding=_snake_case ) _UpperCamelCase : Dict = output.images[0] _UpperCamelCase : Dict = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Any = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = torch_device _UpperCamelCase : int = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) _UpperCamelCase : str = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case ) _UpperCamelCase : List[Any] = output.audios[0] _UpperCamelCase : List[Any] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _UpperCamelCase : Union[str, Any] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Union[str, Any] = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : int = len(UpperCamelCase ) _UpperCamelCase : Optional[Any] = len(matrix[0] ) _UpperCamelCase : str = min(UpperCamelCase ,UpperCamelCase ) for row in range(UpperCamelCase ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 ,UpperCamelCase ): _UpperCamelCase : int = matrix[col][row] / matrix[row][row] for i in range(UpperCamelCase ,UpperCamelCase ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows _UpperCamelCase : str = True for i in range(row + 1 ,UpperCamelCase ): if matrix[i][row] != 0: _UpperCamelCase, _UpperCamelCase : Optional[int] = matrix[i], matrix[row] _UpperCamelCase : Dict = False break if reduce: rank -= 1 for i in range(UpperCamelCase ): _UpperCamelCase : Union[str, Any] = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_longt5""": ["""LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongT5Config""", """LongT5OnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongT5EncoderModel""", """LongT5ForConditionalGeneration""", """LongT5Model""", """LongT5PreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """FlaxLongT5ForConditionalGeneration""", """FlaxLongT5Model""", """FlaxLongT5PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
1
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _UpperCAmelCase : Optional[Any] = """.""" # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) _UpperCAmelCase : Optional[Any] = [ """Assert""", """AssignVariableOp""", """EmptyTensorList""", """MergeV2Checkpoints""", """ReadVariableOp""", """ResourceGather""", """RestoreV2""", """SaveV2""", """ShardedFilename""", """StatefulPartitionedCall""", """StaticRegexFullMatch""", """VarHandleOp""", ] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : int = SavedModel() _UpperCamelCase : List[Any] = [] with open(os.path.join(UpperCamelCase ,'''utils''' ,'''tf_ops''' ,'''onnx.json''' ) ) as f: _UpperCamelCase : Optional[int] = json.load(UpperCamelCase )['''opsets'''] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(UpperCamelCase )] ) with open(UpperCamelCase ,'''rb''' ) as f: saved_model.ParseFromString(f.read() ) _UpperCamelCase : Any = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want _UpperCamelCase : Union[str, Any] = sorted(UpperCamelCase ) _UpperCamelCase : Dict = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(UpperCamelCase ) if strict and len(UpperCamelCase ) > 0: raise Exception(f'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(UpperCamelCase ) > 0: print(f'''Found the following incompatible ops for the opset {opset}:''' ) print(*UpperCamelCase ,sep='''\n''' ) else: print(f'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument("""--saved_model_path""", help="""Path of the saved model to check (the .pb file).""") parser.add_argument( """--opset""", default=12, type=int, help="""The ONNX opset against which the model has to be tested.""" ) parser.add_argument( """--framework""", choices=["""onnx"""], default="""onnx""", help="""Frameworks against which to test the saved model.""" ) parser.add_argument( """--strict""", action="""store_true""", help="""Whether make the checking strict (raise errors) or not (raise warnings)""" ) _UpperCAmelCase : Optional[Any] = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
683
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } _UpperCAmelCase : Any = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : Tuple = DistilBertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> int: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : int = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = do_lower_case _UpperCamelCase : Dict = strip_accents _UpperCamelCase : List[Any] = tokenize_chinese_chars _UpperCamelCase : Tuple = normalizer_class(**_snake_case ) _UpperCamelCase : Dict = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[int]: _UpperCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Optional[Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
1
'''simple docstring''' import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _snake_case , _snake_case=2 , _snake_case=56 , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=99 , _snake_case=32 , _snake_case=2 , _snake_case=2 , _snake_case=7 , _snake_case="gelu_new" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=16 , _snake_case=2 , _snake_case=0.02 , _snake_case=4 , _snake_case="block_sparse" , _snake_case=True , _snake_case=False , _snake_case=2 , _snake_case=3 , ) -> int: _UpperCamelCase : Union[str, Any] = parent _UpperCamelCase : List[Any] = batch_size _UpperCamelCase : Optional[int] = seq_length _UpperCamelCase : List[Any] = is_training _UpperCamelCase : int = use_attention_mask _UpperCamelCase : List[str] = use_token_type_ids _UpperCamelCase : List[str] = use_labels _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : int = hidden_size _UpperCamelCase : Optional[int] = num_hidden_layers _UpperCamelCase : Optional[int] = num_attention_heads _UpperCamelCase : Optional[int] = intermediate_size _UpperCamelCase : Union[str, Any] = hidden_act _UpperCamelCase : List[Any] = hidden_dropout_prob _UpperCamelCase : List[Any] = attention_probs_dropout_prob _UpperCamelCase : int = max_position_embeddings _UpperCamelCase : Union[str, Any] = type_vocab_size _UpperCamelCase : List[str] = type_sequence_label_size _UpperCamelCase : Optional[int] = initializer_range _UpperCamelCase : Tuple = num_choices _UpperCamelCase : List[str] = rescale_embeddings _UpperCamelCase : Union[str, Any] = attention_type _UpperCamelCase : int = use_bias _UpperCamelCase : str = block_size _UpperCamelCase : Optional[Any] = num_random_blocks def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase : List[Any] = None if self.use_attention_mask: _UpperCamelCase : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Optional[int] = None if self.use_token_type_ids: _UpperCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase : List[str] = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_snake_case , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Any = self.prepare_config_and_inputs() _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Union[str, Any] = config_and_inputs _UpperCamelCase : Optional[int] = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Union[str, Any] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) A__ : Optional[Any] = False A__ : Union[str, Any] = False def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : List[Any] = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def _lowercase ( self ) -> Optional[Any]: super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def _lowercase ( self ) -> Optional[int]: super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def _lowercase ( self ) -> Any: super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def _lowercase ( self ) -> Union[str, Any]: super().test_hidden_states_output() @slow def _lowercase ( self ) -> Optional[Any]: for model_class_name in self.all_model_classes: _UpperCamelCase : Optional[Any] = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(_snake_case ) def _lowercase ( self ) -> Dict: if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def _lowercase ( self ) -> Optional[Any]: _UpperCamelCase, _UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCamelCase : List[Any] = self._prepare_for_class(_snake_case , _snake_case ) _UpperCamelCase : Any = model_class(_snake_case ) @jax.jit def model_jitted(_snake_case , _snake_case=None , **_snake_case ): return model(input_ids=_snake_case , attention_mask=_snake_case , **_snake_case ) with self.subTest('''JIT Enabled''' ): _UpperCamelCase : int = model_jitted(**_snake_case ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _UpperCamelCase : int = model_jitted(**_snake_case ).to_tuple() self.assertEqual(len(_snake_case ) , len(_snake_case ) ) for jitted_output, output in zip(_snake_case , _snake_case ): self.assertEqual(jitted_output.shape , output.shape ) def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case=1E-5 , _snake_case="outputs" , _snake_case=None ) -> Any: # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case )
683
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> list: _UpperCamelCase : Any = False while is_sorted is False: # Until all the indices are traversed keep looping _UpperCamelCase : List[str] = True for i in range(0 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Dict = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : int = False for i in range(1 ,len(UpperCamelCase ) - 1 ,2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: _UpperCamelCase, _UpperCamelCase : Optional[Any] = input_list[i + 1], input_list[i] # swapping if elements not in order _UpperCamelCase : Optional[int] = False return input_list if __name__ == "__main__": print("""Enter list to be sorted""") _UpperCAmelCase : Optional[int] = [int(x) for x in input().split()] # inputing elements of the list in one line _UpperCAmelCase : Union[str, Any] = odd_even_sort(input_list) print("""The sorted list is""") print(sorted_list)
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> List[Any]: # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _UpperCamelCase : Optional[Any] = (boundary[1] - boundary[0]) / steps _UpperCamelCase : Optional[int] = boundary[0] _UpperCamelCase : str = boundary[1] _UpperCamelCase : str = make_points(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : int = 0.0 y += (h / 2.0) * f(UpperCamelCase ) for i in x_i: # print(i) y += h * f(UpperCamelCase ) y += (h / 2.0) * f(UpperCamelCase ) return y def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[Any] = a + h while x < (b - h): yield x _UpperCamelCase : Dict = x + h def snake_case__ ( UpperCamelCase ) -> Any: # enter your function here _UpperCamelCase : List[Any] = (x - 0) * (x - 0) return y def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : List[str] = 0.0 # Lower bound of integration _UpperCamelCase : List[str] = 1.0 # Upper bound of integration _UpperCamelCase : Any = 10.0 # define number of steps or resolution _UpperCamelCase : Union[str, Any] = [a, b] # define boundary of integration _UpperCamelCase : List[str] = method_a(UpperCamelCase ,UpperCamelCase ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
683
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : Union[str, Any] = checkpoint _UpperCamelCase : int = {} _UpperCamelCase : int = vae_state_dict['''encoder.conv_in.weight'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_in.bias'''] _UpperCamelCase : Tuple = vae_state_dict['''encoder.conv_out.weight'''] _UpperCamelCase : Any = vae_state_dict['''encoder.conv_out.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''encoder.norm_out.weight'''] _UpperCamelCase : str = vae_state_dict['''encoder.norm_out.bias'''] _UpperCamelCase : str = vae_state_dict['''decoder.conv_in.weight'''] _UpperCamelCase : List[Any] = vae_state_dict['''decoder.conv_in.bias'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.weight'''] _UpperCamelCase : List[str] = vae_state_dict['''decoder.conv_out.bias'''] _UpperCamelCase : int = vae_state_dict['''decoder.norm_out.weight'''] _UpperCamelCase : Dict = vae_state_dict['''decoder.norm_out.bias'''] _UpperCamelCase : Optional[int] = vae_state_dict['''quant_conv.weight'''] _UpperCamelCase : int = vae_state_dict['''quant_conv.bias'''] _UpperCamelCase : List[Any] = vae_state_dict['''post_quant_conv.weight'''] _UpperCamelCase : Optional[int] = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only _UpperCamelCase : Optional[int] = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) _UpperCamelCase : Tuple = { layer_id: [key for key in vae_state_dict if f'''down.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } # Retrieves the keys for the decoder up blocks only _UpperCamelCase : Any = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) _UpperCamelCase : int = { layer_id: [key for key in vae_state_dict if f'''up.{layer_id}''' in key] for layer_id in range(UpperCamelCase ) } for i in range(UpperCamelCase ): _UpperCamelCase : Any = [key for key in down_blocks[i] if f'''down.{i}''' in key and f'''down.{i}.downsample''' not in key] if f'''encoder.down.{i}.downsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Optional[int] = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.weight''' ) _UpperCamelCase : Dict = vae_state_dict.pop( f'''encoder.down.{i}.downsample.conv.bias''' ) _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Union[str, Any] = {'''old''': f'''down.{i}.block''', '''new''': f'''down_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[str] = [key for key in vae_state_dict if '''encoder.mid.block''' in key] _UpperCamelCase : Tuple = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : Optional[int] = [key for key in mid_resnets if f'''encoder.mid.block_{i}''' in key] _UpperCamelCase : List[str] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Tuple = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] _UpperCamelCase : List[str] = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) for i in range(UpperCamelCase ): _UpperCamelCase : Union[str, Any] = num_up_blocks - 1 - i _UpperCamelCase : Optional[int] = [ key for key in up_blocks[block_id] if f'''up.{block_id}''' in key and f'''up.{block_id}.upsample''' not in key ] if f'''decoder.up.{block_id}.upsample.conv.weight''' in vae_state_dict: _UpperCamelCase : Tuple = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.weight''' ] _UpperCamelCase : Any = vae_state_dict[ f'''decoder.up.{block_id}.upsample.conv.bias''' ] _UpperCamelCase : Any = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Any = {'''old''': f'''up.{block_id}.block''', '''new''': f'''up_blocks.{i}.resnets'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : List[Any] = [key for key in vae_state_dict if '''decoder.mid.block''' in key] _UpperCamelCase : Optional[Any] = 2 for i in range(1 ,num_mid_res_blocks + 1 ): _UpperCamelCase : int = [key for key in mid_resnets if f'''decoder.mid.block_{i}''' in key] _UpperCamelCase : Optional[int] = renew_vae_resnet_paths(UpperCamelCase ) _UpperCamelCase : Optional[Any] = {'''old''': f'''mid.block_{i}''', '''new''': f'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) _UpperCamelCase : Tuple = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] _UpperCamelCase : Tuple = renew_vae_attention_paths(UpperCamelCase ) _UpperCamelCase : Dict = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,additional_replacements=[meta_path] ,config=UpperCamelCase ) conv_attn_to_linear(UpperCamelCase ) return new_checkpoint def snake_case__ ( UpperCamelCase ,UpperCamelCase ,) -> List[str]: # Only support V1 _UpperCamelCase : Tuple = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) _UpperCamelCase : List[Any] = io.BytesIO(r.content ) _UpperCamelCase : Optional[int] = OmegaConf.load(UpperCamelCase ) _UpperCamelCase : str = 5_12 _UpperCamelCase : int = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open _UpperCamelCase : str = {} with safe_open(UpperCamelCase ,framework='''pt''' ,device='''cpu''' ) as f: for key in f.keys(): _UpperCamelCase : Union[str, Any] = f.get_tensor(UpperCamelCase ) else: _UpperCamelCase : str = torch.load(UpperCamelCase ,map_location=UpperCamelCase )['''state_dict'''] # Convert the VAE model. _UpperCamelCase : Dict = create_vae_diffusers_config(UpperCamelCase ,image_size=UpperCamelCase ) _UpperCamelCase : str = custom_convert_ldm_vae_checkpoint(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Dict = AutoencoderKL(**UpperCamelCase ) vae.load_state_dict(UpperCamelCase ) vae.save_pretrained(UpperCamelCase ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") _UpperCAmelCase : int = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
683
1
'''simple docstring''' import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def snake_case__ ( UpperCamelCase ,UpperCamelCase ,**UpperCamelCase ) -> int: _UpperCamelCase : Tuple = AutoConfig.from_pretrained(UpperCamelCase ,**UpperCamelCase ) _UpperCamelCase : Optional[int] = AutoModelForSeqaSeqLM.from_config(UpperCamelCase ) model.save_pretrained(UpperCamelCase ) AutoTokenizer.from_pretrained(UpperCamelCase ).save_pretrained(UpperCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
683
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = ['image_processor', 'tokenizer'] A__ : Dict = 'CLIPImageProcessor' A__ : str = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self , _snake_case=None , _snake_case=None , **_snake_case ) -> List[Any]: _UpperCamelCase : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) _UpperCamelCase : Optional[Any] = kwargs.pop('''feature_extractor''' ) _UpperCamelCase : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_snake_case , _snake_case ) def __call__( self , _snake_case=None , _snake_case=None , _snake_case=None , **_snake_case ) -> Dict: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: _UpperCamelCase : List[str] = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: _UpperCamelCase : str = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: _UpperCamelCase : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Tuple: return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> Any: return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def _lowercase ( self ) -> int: _UpperCamelCase : Optional[int] = self.tokenizer.model_input_names _UpperCamelCase : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
683
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: _UpperCAmelCase : Tuple = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : str = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Tuple = { """vocab_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model""", }, """tokenizer_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json""", }, } _UpperCAmelCase : List[str] = { """xlnet-base-cased""": None, """xlnet-large-cased""": None, } _UpperCAmelCase : List[Any] = """▁""" # Segments (not really needed) _UpperCAmelCase : str = 0 _UpperCAmelCase : Optional[Any] = 1 _UpperCAmelCase : Optional[Any] = 2 _UpperCAmelCase : str = 3 _UpperCAmelCase : str = 4 class UpperCAmelCase ( a_ ): """simple docstring""" A__ : int = VOCAB_FILES_NAMES A__ : str = PRETRAINED_VOCAB_FILES_MAP A__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : List[str] = 'left' A__ : List[Any] = XLNetTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=False , _snake_case=True , _snake_case=False , _snake_case="<s>" , _snake_case="</s>" , _snake_case="<unk>" , _snake_case="<sep>" , _snake_case="<pad>" , _snake_case="<cls>" , _snake_case="<mask>" , _snake_case=["<eop>", "<eod>"] , **_snake_case , ) -> List[Any]: # Mask token behave like a normal word, i.e. include the space before it _UpperCamelCase : Optional[int] = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else mask_token super().__init__( vocab_file=_snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , remove_space=_snake_case , keep_accents=_snake_case , bos_token=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , additional_special_tokens=_snake_case , **_snake_case , ) _UpperCamelCase : str = 3 _UpperCamelCase : Any = do_lower_case _UpperCamelCase : int = remove_space _UpperCamelCase : List[str] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : List[Any] = False if not self.vocab_file else True def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Optional[int] = [self.sep_token_id] _UpperCamelCase : Dict = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_snake_case ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase : Optional[int] = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_snake_case ): copyfile(self.vocab_file , _snake_case ) return (out_vocab_file,)
683
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _UpperCAmelCase : Union[str, Any] = (720, 1280) # Height, Width _UpperCAmelCase : str = (0.4, 0.6) # if height or width lower than this scale, drop it. _UpperCAmelCase : Optional[Any] = 1 / 100 _UpperCAmelCase : Optional[Any] = """""" _UpperCAmelCase : int = """""" _UpperCAmelCase : Union[str, Any] = """""" _UpperCAmelCase : List[Any] = 250 def snake_case__ ( ) -> None: _UpperCamelCase, _UpperCamelCase : List[Any] = get_dataset(UpperCamelCase ,UpperCamelCase ) for index in range(UpperCamelCase ): _UpperCamelCase : List[str] = random.sample(range(len(UpperCamelCase ) ) ,4 ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[str] = update_image_and_anno( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,filter_scale=UpperCamelCase ,) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _UpperCamelCase : List[str] = random_chars(32 ) _UpperCamelCase : List[str] = path.split(os.sep )[-1].rsplit('''.''' ,1 )[0] _UpperCamelCase : Any = f'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(f'''{file_root}.jpg''' ,UpperCamelCase ,[cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) _UpperCamelCase : Any = [] for anno in new_annos: _UpperCamelCase : List[Any] = anno[3] - anno[1] _UpperCamelCase : int = anno[4] - anno[2] _UpperCamelCase : int = anno[1] + width / 2 _UpperCamelCase : int = anno[2] + height / 2 _UpperCamelCase : Optional[Any] = f'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(UpperCamelCase ) with open(f'''{file_root}.txt''' ,'''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> tuple[list, list]: _UpperCamelCase : List[str] = [] _UpperCamelCase : Union[str, Any] = [] for label_file in glob.glob(os.path.join(UpperCamelCase ,'''*.txt''' ) ): _UpperCamelCase : int = label_file.split(os.sep )[-1].rsplit('''.''' ,1 )[0] with open(UpperCamelCase ) as in_file: _UpperCamelCase : Dict = in_file.readlines() _UpperCamelCase : Tuple = os.path.join(UpperCamelCase ,f'''{label_name}.jpg''' ) _UpperCamelCase : Tuple = [] for obj_list in obj_lists: _UpperCamelCase : List[Any] = obj_list.rstrip('''\n''' ).split(''' ''' ) _UpperCamelCase : Tuple = float(obj[1] ) - float(obj[3] ) / 2 _UpperCamelCase : Any = float(obj[2] ) - float(obj[4] ) / 2 _UpperCamelCase : Tuple = float(obj[1] ) + float(obj[3] ) / 2 _UpperCamelCase : List[Any] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(UpperCamelCase ) labels.append(UpperCamelCase ) return img_paths, labels def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = 0.0 ,) -> tuple[list, list, str]: _UpperCamelCase : Optional[int] = np.zeros([output_size[0], output_size[1], 3] ,dtype=np.uinta ) _UpperCamelCase : str = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _UpperCamelCase : Dict = int(scale_x * output_size[1] ) _UpperCamelCase : Dict = int(scale_y * output_size[0] ) _UpperCamelCase : int = [] _UpperCamelCase : Union[str, Any] = [] for i, index in enumerate(UpperCamelCase ): _UpperCamelCase : Optional[int] = all_img_list[index] path_list.append(UpperCamelCase ) _UpperCamelCase : str = all_annos[index] _UpperCamelCase : Tuple = cva.imread(UpperCamelCase ) if i == 0: # top-left _UpperCamelCase : Any = cva.resize(UpperCamelCase ,(divid_point_x, divid_point_y) ) _UpperCamelCase : Any = img for bbox in img_annos: _UpperCamelCase : List[Any] = bbox[1] * scale_x _UpperCamelCase : Dict = bbox[2] * scale_y _UpperCamelCase : Any = bbox[3] * scale_x _UpperCamelCase : Any = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _UpperCamelCase : Union[str, Any] = cva.resize(UpperCamelCase ,(output_size[1] - divid_point_x, divid_point_y) ) _UpperCamelCase : List[Any] = img for bbox in img_annos: _UpperCamelCase : Any = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Optional[Any] = bbox[2] * scale_y _UpperCamelCase : Any = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : Optional[int] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _UpperCamelCase : Dict = cva.resize(UpperCamelCase ,(divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : List[str] = img for bbox in img_annos: _UpperCamelCase : int = bbox[1] * scale_x _UpperCamelCase : Optional[Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : int = bbox[3] * scale_x _UpperCamelCase : Any = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _UpperCamelCase : Dict = cva.resize( UpperCamelCase ,(output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _UpperCamelCase : Union[str, Any] = img for bbox in img_annos: _UpperCamelCase : Optional[int] = scale_x + bbox[1] * (1 - scale_x) _UpperCamelCase : Union[str, Any] = scale_y + bbox[2] * (1 - scale_y) _UpperCamelCase : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) _UpperCamelCase : List[str] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _UpperCamelCase : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def snake_case__ ( UpperCamelCase ) -> str: assert number_char > 1, "The number of character should greater than 1" _UpperCamelCase : Tuple = ascii_lowercase + digits return "".join(random.choice(UpperCamelCase ) for _ in range(UpperCamelCase ) ) if __name__ == "__main__": main() print("""DONE ✅""")
683
1
'''simple docstring''' import functools def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = len(UpperCamelCase ) _UpperCamelCase : str = len(UpperCamelCase ) @functools.cache def min_distance(UpperCamelCase ,UpperCamelCase ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa _UpperCamelCase : Tuple = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 ,UpperCamelCase ) ,1 + min_distance(UpperCamelCase ,indexa + 1 ) ,diff + min_distance(indexa + 1 ,indexa + 1 ) ,) return min_distance(0 ,0 ) if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class UpperCAmelCase ( a_ ): """simple docstring""" @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Union[str, Any] = EncoderDecoderModel.from_encoder_decoder_pretrained('''prajjwal1/bert-tiny''' , '''prajjwal1/bert-tiny''' ) _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained('''bert-base-uncased''' ) _UpperCamelCase : Optional[Any] = bertabert.config.encoder.vocab_size _UpperCamelCase : List[str] = tokenizer.sep_token_id _UpperCamelCase : List[str] = tokenizer.cls_token_id _UpperCamelCase : Optional[Any] = 128 _UpperCamelCase : int = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''train[:1%]''' ) _UpperCamelCase : Dict = datasets.load_dataset('''cnn_dailymail''' , '''3.0.0''' , split='''validation[:1%]''' ) _UpperCamelCase : Dict = train_dataset.select(range(32 ) ) _UpperCamelCase : Tuple = val_dataset.select(range(16 ) ) _UpperCamelCase : Union[str, Any] = 4 def _map_to_encoder_decoder_inputs(_snake_case ): # Tokenizer will automatically set [BOS] <text> [EOS] _UpperCamelCase : Optional[Any] = tokenizer(batch['''article'''] , padding='''max_length''' , truncation=_snake_case , max_length=512 ) _UpperCamelCase : Optional[int] = tokenizer(batch['''highlights'''] , padding='''max_length''' , truncation=_snake_case , max_length=128 ) _UpperCamelCase : str = inputs.input_ids _UpperCamelCase : Union[str, Any] = inputs.attention_mask _UpperCamelCase : str = outputs.input_ids _UpperCamelCase : str = outputs.input_ids.copy() _UpperCamelCase : Tuple = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['''labels'''] ] _UpperCamelCase : Union[str, Any] = outputs.attention_mask assert all(len(_snake_case ) == 512 for x in inputs.input_ids ) assert all(len(_snake_case ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_snake_case ): _UpperCamelCase : Dict = pred.label_ids _UpperCamelCase : Optional[int] = pred.predictions # all unnecessary tokens are removed _UpperCamelCase : Any = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : Dict = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) _UpperCamelCase : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_snake_case ) )] ) / len(_snake_case ) return {"accuracy": accuracy} # map train dataset _UpperCamelCase : Optional[Any] = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) train_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) # same for validation dataset _UpperCamelCase : List[Any] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_snake_case , batch_size=_snake_case , remove_columns=['''article''', '''highlights'''] , ) val_dataset.set_format( type='''torch''' , columns=['''input_ids''', '''attention_mask''', '''decoder_input_ids''', '''decoder_attention_mask''', '''labels'''] , ) _UpperCamelCase : Optional[int] = self.get_auto_remove_tmp_dir() _UpperCamelCase : Union[str, Any] = SeqaSeqTrainingArguments( output_dir=_snake_case , per_device_train_batch_size=_snake_case , per_device_eval_batch_size=_snake_case , predict_with_generate=_snake_case , evaluation_strategy='''steps''' , do_train=_snake_case , do_eval=_snake_case , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _UpperCamelCase : Optional[int] = SeqaSeqTrainer( model=_snake_case , args=_snake_case , compute_metrics=_compute_metrics , train_dataset=_snake_case , eval_dataset=_snake_case , tokenizer=_snake_case , ) # start training trainer.train()
683
1
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging _UpperCAmelCase : str = logging.get_logger(__name__) class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = ['input_features', 'is_longer'] def __init__( self , _snake_case=64 , _snake_case=48000 , _snake_case=480 , _snake_case=10 , _snake_case=1024 , _snake_case=0.0 , _snake_case=False , _snake_case = 0 , _snake_case = 14000 , _snake_case = None , _snake_case = "fusion" , _snake_case = "repeatpad" , **_snake_case , ) -> Optional[int]: super().__init__( feature_size=_snake_case , sampling_rate=_snake_case , padding_value=_snake_case , return_attention_mask=_snake_case , **_snake_case , ) _UpperCamelCase : Optional[int] = top_db _UpperCamelCase : Tuple = truncation _UpperCamelCase : Any = padding _UpperCamelCase : Optional[Any] = fft_window_size _UpperCamelCase : Optional[int] = (fft_window_size >> 1) + 1 _UpperCamelCase : List[Any] = hop_length _UpperCamelCase : List[Any] = max_length_s _UpperCamelCase : Tuple = max_length_s * sampling_rate _UpperCamelCase : str = sampling_rate _UpperCamelCase : List[Any] = frequency_min _UpperCamelCase : int = frequency_max _UpperCamelCase : Union[str, Any] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_snake_case , min_frequency=_snake_case , max_frequency=_snake_case , sampling_rate=_snake_case , norm=_snake_case , mel_scale='''htk''' , ) _UpperCamelCase : Dict = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_snake_case , min_frequency=_snake_case , max_frequency=_snake_case , sampling_rate=_snake_case , norm='''slaney''' , mel_scale='''slaney''' , ) def _lowercase ( self ) -> Dict[str, Any]: _UpperCamelCase : Tuple = copy.deepcopy(self.__dict__ ) _UpperCamelCase : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> np.ndarray: _UpperCamelCase : Optional[int] = spectrogram( _snake_case , window_function(self.fft_window_size , '''hann''' ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=_snake_case , log_mel='''dB''' , ) return log_mel_spectrogram.T def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : List[str] = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk _UpperCamelCase : str = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk _UpperCamelCase : Tuple = [0] # randomly choose index for each part _UpperCamelCase : List[str] = np.random.choice(ranges[0] ) _UpperCamelCase : Dict = np.random.choice(ranges[1] ) _UpperCamelCase : str = np.random.choice(ranges[2] ) _UpperCamelCase : Optional[Any] = mel[idx_front : idx_front + chunk_frames, :] _UpperCamelCase : List[Any] = mel[idx_middle : idx_middle + chunk_frames, :] _UpperCamelCase : Tuple = mel[idx_back : idx_back + chunk_frames, :] _UpperCamelCase : Tuple = torch.tensor(mel[None, None, :] ) _UpperCamelCase : int = torch.nn.functional.interpolate( _snake_case , size=[chunk_frames, 64] , mode='''bilinear''' , align_corners=_snake_case ) _UpperCamelCase : Optional[Any] = mel_shrink[0][0].numpy() _UpperCamelCase : str = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def _lowercase ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> np.array: if waveform.shape[0] > max_length: if truncation == "rand_trunc": _UpperCamelCase : Tuple = True # random crop to max_length (for compatibility) -> this should be handled by self.pad _UpperCamelCase : int = len(_snake_case ) - max_length _UpperCamelCase : int = np.random.randint(0 , overflow + 1 ) _UpperCamelCase : Optional[Any] = waveform[idx : idx + max_length] _UpperCamelCase : int = self._np_extract_fbank_features(_snake_case , self.mel_filters_slaney )[None, :] elif truncation == "fusion": _UpperCamelCase : Any = self._np_extract_fbank_features(_snake_case , self.mel_filters ) _UpperCamelCase : Optional[Any] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed _UpperCamelCase : Dict = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. _UpperCamelCase : Any = np.stack([mel, mel, mel, mel] , axis=0 ) _UpperCamelCase : str = False else: _UpperCamelCase : Any = self._random_mel_fusion(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = True else: raise NotImplementedError(F'''data_truncating {truncation} not implemented''' ) else: _UpperCamelCase : Tuple = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": _UpperCamelCase : Tuple = int(max_length / len(_snake_case ) ) _UpperCamelCase : List[Any] = np.stack(np.tile(_snake_case , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": _UpperCamelCase : Optional[int] = int(max_length / len(_snake_case ) ) _UpperCamelCase : List[Any] = np.stack(np.tile(_snake_case , _snake_case ) ) _UpperCamelCase : Optional[int] = np.pad(_snake_case , (0, max_length - waveform.shape[0]) , mode='''constant''' , constant_values=0 ) if truncation == "fusion": _UpperCamelCase : Dict = self._np_extract_fbank_features(_snake_case , self.mel_filters ) _UpperCamelCase : int = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: _UpperCamelCase : Optional[int] = self._np_extract_fbank_features(_snake_case , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self , _snake_case , _snake_case = None , _snake_case = None , _snake_case = None , _snake_case = None , _snake_case = None , **_snake_case , ) -> BatchFeature: _UpperCamelCase : Any = truncation if truncation is not None else self.truncation _UpperCamelCase : Tuple = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' F''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' F''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) _UpperCamelCase : Tuple = isinstance(_snake_case , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) _UpperCamelCase : Dict = is_batched_numpy or ( isinstance(_snake_case , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _UpperCamelCase : Optional[Any] = [np.asarray(_snake_case , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_snake_case , np.ndarray ): _UpperCamelCase : List[str] = np.asarray(_snake_case , dtype=np.floataa ) elif isinstance(_snake_case , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _UpperCamelCase : Tuple = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _UpperCamelCase : Tuple = [np.asarray(_snake_case )] # convert to mel spectrogram, truncate and pad if needed. _UpperCamelCase : Optional[int] = [ self._get_input_mel(_snake_case , max_length if max_length else self.nb_max_samples , _snake_case , _snake_case ) for waveform in raw_speech ] _UpperCamelCase : str = [] _UpperCamelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(_snake_case ) is_longer.append(_snake_case ) if truncation == "fusion" and sum(_snake_case ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer _UpperCamelCase : Optional[int] = np.random.randint(0 , len(_snake_case ) ) _UpperCamelCase : Any = True if isinstance(input_mel[0] , _snake_case ): _UpperCamelCase : str = [np.asarray(_snake_case , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool _UpperCamelCase : Optional[Any] = [[longer] for longer in is_longer] _UpperCamelCase : List[str] = {'''input_features''': input_mel, '''is_longer''': is_longer} _UpperCamelCase : str = BatchFeature(_snake_case ) if return_tensors is not None: _UpperCamelCase : Union[str, Any] = input_features.convert_to_tensors(_snake_case ) return input_features
683
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def snake_case__ ( UpperCamelCase=None ) -> Optional[int]: if subparsers is not None: _UpperCamelCase : Dict = subparsers.add_parser('''env''' ) else: _UpperCamelCase : Tuple = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''' ,default=UpperCamelCase ,help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase ) return parser def snake_case__ ( UpperCamelCase ) -> Any: _UpperCamelCase : int = torch.__version__ _UpperCamelCase : int = torch.cuda.is_available() _UpperCamelCase : List[str] = is_xpu_available() _UpperCamelCase : Dict = is_npu_available() _UpperCamelCase : Optional[Any] = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(UpperCamelCase ): _UpperCamelCase : List[str] = load_config_from_file(args.config_file ).to_dict() _UpperCamelCase : List[Any] = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': f'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(UpperCamelCase ), '''PyTorch NPU available''': str(UpperCamelCase ), '''System RAM''': f'''{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB''', } if pt_cuda_available: _UpperCamelCase : int = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([f'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) _UpperCamelCase : Union[str, Any] = ( '''\n'''.join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(UpperCamelCase ,UpperCamelCase ) else f'''\t{accelerate_config}''' ) print(UpperCamelCase ) _UpperCamelCase : str = accelerate_config return info def snake_case__ ( ) -> int: _UpperCamelCase : str = env_command_parser() _UpperCamelCase : Any = parser.parse_args() env_command(UpperCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
683
1
'''simple docstring''' import itertools import string from collections.abc import Generator, Iterable def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Generator[tuple[str, ...], None, None]: _UpperCamelCase : int = iter(UpperCamelCase ) while True: _UpperCamelCase : Union[str, Any] = tuple(itertools.islice(UpperCamelCase ,UpperCamelCase ) ) if not chunk: return yield chunk def snake_case__ ( UpperCamelCase ) -> str: _UpperCamelCase : Union[str, Any] = ''''''.join([c.upper() for c in dirty if c in string.ascii_letters] ) _UpperCamelCase : int = '''''' if len(UpperCamelCase ) < 2: return dirty for i in range(len(UpperCamelCase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(UpperCamelCase ) & 1: clean += "X" return clean def snake_case__ ( UpperCamelCase ) -> list[str]: # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) _UpperCamelCase : int = '''ABCDEFGHIKLMNOPQRSTUVWXYZ''' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _UpperCamelCase : Tuple = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(UpperCamelCase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(UpperCamelCase ) return table def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> str: _UpperCamelCase : Optional[int] = generate_table(UpperCamelCase ) _UpperCamelCase : List[str] = prepare_input(UpperCamelCase ) _UpperCamelCase : Dict = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase ,2 ): _UpperCamelCase, _UpperCamelCase : Tuple = divmod(table.index(UpperCamelCase ) ,5 ) _UpperCamelCase, _UpperCamelCase : int = divmod(table.index(UpperCamelCase ) ,5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> str: _UpperCamelCase : Dict = generate_table(UpperCamelCase ) _UpperCamelCase : Dict = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase ,2 ): _UpperCamelCase, _UpperCamelCase : List[Any] = divmod(table.index(UpperCamelCase ) ,5 ) _UpperCamelCase, _UpperCamelCase : List[Any] = divmod(table.index(UpperCamelCase ) ,5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
683
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : List[Any] = logging.get_logger(__name__) def snake_case__ ( UpperCamelCase ) -> Tuple: _UpperCamelCase : str = '''huggingface/label-files''' _UpperCamelCase : Optional[Any] = '''imagenet-1k-id2label.json''' _UpperCamelCase : Optional[int] = json.load(open(hf_hub_download(UpperCamelCase ,UpperCamelCase ,repo_type='''dataset''' ) ,'''r''' ) ) _UpperCamelCase : Optional[int] = {int(UpperCamelCase ): v for k, v in idalabel.items()} _UpperCamelCase : Dict = {v: k for k, v in idalabel.items()} _UpperCamelCase : Optional[Any] = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" _UpperCamelCase : Union[str, Any] = BitConfig( conv_layer=UpperCamelCase ,num_labels=10_00 ,idalabel=UpperCamelCase ,labelaid=UpperCamelCase ,) return config def snake_case__ ( UpperCamelCase ) -> str: if "stem.conv" in name: _UpperCamelCase : Any = name.replace('''stem.conv''' ,'''bit.embedder.convolution''' ) if "blocks" in name: _UpperCamelCase : Union[str, Any] = name.replace('''blocks''' ,'''layers''' ) if "head.fc" in name: _UpperCamelCase : Optional[Any] = name.replace('''head.fc''' ,'''classifier.1''' ) if name.startswith('''norm''' ): _UpperCamelCase : Any = '''bit.''' + name if "bit" not in name and "classifier" not in name: _UpperCamelCase : List[Any] = '''bit.encoder.''' + name return name def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase : List[str] = Image.open(requests.get(UpperCamelCase ,stream=UpperCamelCase ).raw ) return im @torch.no_grad() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[Any]: _UpperCamelCase : str = get_config(UpperCamelCase ) # load original model from timm _UpperCamelCase : int = create_model(UpperCamelCase ,pretrained=UpperCamelCase ) timm_model.eval() # load state_dict of original model _UpperCamelCase : int = timm_model.state_dict() for key in state_dict.copy().keys(): _UpperCamelCase : int = state_dict.pop(UpperCamelCase ) _UpperCamelCase : Any = val.squeeze() if '''head''' in key else val # load HuggingFace model _UpperCamelCase : List[str] = BitForImageClassification(UpperCamelCase ) model.eval() model.load_state_dict(UpperCamelCase ) # create image processor _UpperCamelCase : Optional[int] = create_transform(**resolve_data_config({} ,model=UpperCamelCase ) ) _UpperCamelCase : Any = transform.transforms _UpperCamelCase : List[str] = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } _UpperCamelCase : List[str] = BitImageProcessor( do_resize=UpperCamelCase ,size={'''shortest_edge''': timm_transforms[0].size} ,resample=pillow_resamplings[timm_transforms[0].interpolation.value] ,do_center_crop=UpperCamelCase ,crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]} ,do_normalize=UpperCamelCase ,image_mean=timm_transforms[-1].mean.tolist() ,image_std=timm_transforms[-1].std.tolist() ,) _UpperCamelCase : str = prepare_img() _UpperCamelCase : Dict = transform(UpperCamelCase ).unsqueeze(0 ) _UpperCamelCase : Dict = processor(UpperCamelCase ,return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(UpperCamelCase ,UpperCamelCase ) # verify logits with torch.no_grad(): _UpperCamelCase : Optional[int] = model(UpperCamelCase ) _UpperCamelCase : Optional[int] = outputs.logits print('''Logits:''' ,logits[0, :3] ) print('''Predicted class:''' ,model.config.idalabel[logits.argmax(-1 ).item()] ) _UpperCamelCase : List[Any] = timm_model(UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase ,outputs.logits ,atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": _UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) _UpperCAmelCase : Optional[int] = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
683
1
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[int] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Dict = XGLMTokenizer A__ : Any = XGLMTokenizerFast A__ : Optional[int] = True A__ : Any = True def _lowercase ( self ) -> Optional[int]: super().setUp() # We have a SentencePiece fixture for testing _UpperCamelCase : Optional[Any] = XGLMTokenizer(_snake_case , keep_accents=_snake_case ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : List[Any] = '''<pad>''' _UpperCamelCase : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_snake_case ) , _snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_snake_case ) , _snake_case ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(len(_snake_case ) , 1008 ) def _lowercase ( self ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = XGLMTokenizer(_snake_case , keep_accents=_snake_case ) _UpperCamelCase : Dict = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_snake_case , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_snake_case ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _UpperCamelCase : Union[str, Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _snake_case , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _UpperCamelCase : Optional[int] = tokenizer.convert_tokens_to_ids(_snake_case ) self.assertListEqual( _snake_case , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _UpperCamelCase : str = tokenizer.convert_ids_to_tokens(_snake_case ) self.assertListEqual( _snake_case , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def _lowercase ( self ) -> Optional[Any]: return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def _lowercase ( self ) -> str: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(_snake_case , f.name ) _UpperCamelCase : int = XGLMTokenizer(f.name , keep_accents=_snake_case ) _UpperCamelCase : List[Any] = pickle.dumps(_snake_case ) pickle.loads(_snake_case ) def _lowercase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return _UpperCamelCase : str = self.get_tokenizer() _UpperCamelCase : int = self.get_rust_tokenizer() _UpperCamelCase : Optional[Any] = '''I was born in 92000, and this is falsé.''' _UpperCamelCase : Union[str, Any] = tokenizer.tokenize(_snake_case ) _UpperCamelCase : str = rust_tokenizer.tokenize(_snake_case ) self.assertListEqual(_snake_case , _snake_case ) _UpperCamelCase : Tuple = tokenizer.encode(_snake_case , add_special_tokens=_snake_case ) _UpperCamelCase : Dict = rust_tokenizer.encode(_snake_case , add_special_tokens=_snake_case ) self.assertListEqual(_snake_case , _snake_case ) _UpperCamelCase : List[str] = self.get_rust_tokenizer() _UpperCamelCase : Dict = tokenizer.encode(_snake_case ) _UpperCamelCase : Optional[Any] = rust_tokenizer.encode(_snake_case ) self.assertListEqual(_snake_case , _snake_case ) @slow def _lowercase ( self ) -> int: _UpperCamelCase : Union[str, Any] = '''Hello World!''' _UpperCamelCase : Dict = [2, 31227, 4447, 35] self.assertListEqual(_snake_case , self.big_tokenizer.encode(_snake_case ) ) @slow def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off _UpperCamelCase : Dict = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 71630, 28085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 13675, 377, 652, 7580, 10341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 202277, 17892, 33, 60, 87, 4, 3234, 157, 61, 2667, 52376, 19, 88, 23, 735] # fmt: on self.assertListEqual(_snake_case , self.big_tokenizer.encode(_snake_case ) ) @slow def _lowercase ( self ) -> Union[str, Any]: # fmt: off _UpperCamelCase : Dict = { '''input_ids''': [[2, 108825, 1163, 15, 88010, 473, 15898, 157, 13672, 1857, 312, 8, 238021, 1163, 53, 13672, 1857, 312, 8, 53283, 182396, 8, 18566, 16, 36733, 4101, 8, 230, 244017, 122553, 7, 15, 132597, 4, 293, 12511, 7610, 4, 3414, 132597, 9, 4, 32361, 362, 4, 734, 28512, 32569, 18, 4, 32361, 26096, 14982, 73, 18715, 21433, 235261, 15, 492, 12427, 16, 53, 18715, 21433, 65454, 15, 23659, 563, 16, 278, 597, 2843, 595, 7931, 182396, 64186, 22, 886, 595, 132981, 53, 25540, 3449, 43982, 39901, 5951, 878, 330, 4, 27694, 80269, 312, 53, 6517, 11780, 611, 20408, 5], [2, 6, 132597, 67, 42897, 33, 592, 8, 163729, 25540, 361, 136997, 109514, 173230, 7, 501, 60, 102913, 196, 5631, 235, 63243, 473, 6, 231757, 74, 5277, 7905, 53, 3095, 37317, 22, 454, 183874, 5], [2, 268, 31298, 46530, 6, 132935, 43831, 7, 597, 32, 24, 3688, 9865, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_snake_case , model_name='''facebook/xglm-564M''' , padding=_snake_case , )
683
'''simple docstring''' _UpperCAmelCase : Any = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100000)] def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : Any = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_00_00] number //= 10_00_00 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _UpperCAmelCase : list[bool | None] = [None] * 10000000 _UpperCAmelCase : str = True _UpperCAmelCase : Tuple = False def snake_case__ ( UpperCamelCase ) -> bool: if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _UpperCamelCase : List[str] = chain(next_number(UpperCamelCase ) ) _UpperCamelCase : Tuple = number_chain while number < 10_00_00_00: _UpperCamelCase : int = number_chain number *= 10 return number_chain def snake_case__ ( UpperCamelCase = 10_00_00_00 ) -> int: for i in range(1 ,UpperCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
683
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Tuple = 'microsoft/speecht5_tts' A__ : Dict = ( 'This is a tool that reads an English text out loud. It takes an input named `text` which should contain the ' 'text to read (in English) and returns a waveform object containing the sound.' ) A__ : str = 'text_reader' A__ : List[Any] = SpeechTaProcessor A__ : Any = SpeechTaForTextToSpeech A__ : List[Any] = SpeechTaHifiGan A__ : str = ['text'] A__ : str = ['audio'] def _lowercase ( self ) -> str: if self.post_processor is None: _UpperCamelCase : int = '''microsoft/speecht5_hifigan''' super().setup() def _lowercase ( self , _snake_case , _snake_case=None ) -> Dict: _UpperCamelCase : Tuple = self.pre_processor(text=_snake_case , return_tensors='''pt''' , truncation=_snake_case ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError('''Datasets needs to be installed if not passing speaker embeddings.''' ) _UpperCamelCase : Optional[int] = load_dataset('''Matthijs/cmu-arctic-xvectors''' , split='''validation''' ) _UpperCamelCase : str = torch.tensor(embeddings_dataset[7305]['''xvector'''] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def _lowercase ( self , _snake_case ) -> List[str]: with torch.no_grad(): return self.model.generate_speech(**_snake_case ) def _lowercase ( self , _snake_case ) -> Optional[Any]: with torch.no_grad(): return self.post_processor(_snake_case ).cpu().detach()
683
'''simple docstring''' _UpperCAmelCase : str = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _UpperCAmelCase : List[str] = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> str: assert len(str(UpperCamelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _UpperCamelCase : Any = year // 1_00 _UpperCamelCase : List[Any] = (5 * (century % 4) + 2) % 7 _UpperCamelCase : Tuple = year % 1_00 _UpperCamelCase : Optional[int] = centurian % 12 _UpperCamelCase : Tuple = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _UpperCamelCase : List[Any] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_00) == 0) else DOOMSDAY_LEAP[month - 1] ) _UpperCamelCase : Optional[int] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> bool: return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
683
'''simple docstring''' import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : """simple docstring""" @staticmethod def _lowercase ( *_snake_case , **_snake_case ) -> str: pass @is_pipeline_test @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" A__ : Tuple = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def _lowercase ( self , _snake_case , _snake_case , _snake_case ) -> Optional[Any]: _UpperCamelCase : int = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Any = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def _lowercase ( self , _snake_case , _snake_case ) -> List[str]: _UpperCamelCase : int = vqa_pipeline(_snake_case , top_k=1 ) self.assertEqual( _snake_case , [ [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}], ] , ) @require_torch def _lowercase ( self ) -> Tuple: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[int] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question='''How many cats are there?''' , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) _UpperCamelCase : List[Any] = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( _snake_case , [{'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}, {'''score''': ANY(_snake_case ), '''answer''': ANY(_snake_case )}] ) @slow @require_torch def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Any = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''' ) _UpperCamelCase : Dict = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _UpperCamelCase : Optional[Any] = '''How many cats are there?''' _UpperCamelCase : str = vqa_pipeline(image=_snake_case , question=_snake_case , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : str = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}] ) _UpperCamelCase : Dict = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(_snake_case , decimals=4 ) , [[{'''score''': 0.8_799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''' ) def _lowercase ( self ) -> List[Any]: pass
683
1
'''simple docstring''' import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = (IPNDMScheduler,) A__ : List[str] = (('num_inference_steps', 50),) def _lowercase ( self , **_snake_case ) -> Tuple: _UpperCamelCase : List[str] = {'''num_train_timesteps''': 1000} config.update(**_snake_case ) return config def _lowercase ( self , _snake_case=0 , **_snake_case ) -> Tuple: _UpperCamelCase : Optional[Any] = dict(self.forward_default_kwargs ) _UpperCamelCase : Dict = kwargs.pop('''num_inference_steps''' , _snake_case ) _UpperCamelCase : str = self.dummy_sample _UpperCamelCase : str = 0.1 * sample _UpperCamelCase : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _UpperCamelCase : Optional[Any] = self.get_scheduler_config(**_snake_case ) _UpperCamelCase : str = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals _UpperCamelCase : Union[str, Any] = dummy_past_residuals[:] if time_step is None: _UpperCamelCase : Any = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) _UpperCamelCase : List[str] = scheduler_class.from_pretrained(_snake_case ) new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals _UpperCamelCase : Dict = dummy_past_residuals[:] _UpperCamelCase : Union[str, Any] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : Optional[int] = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _UpperCamelCase : Optional[int] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : Any = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self ) -> List[str]: pass def _lowercase ( self , _snake_case=0 , **_snake_case ) -> Tuple: _UpperCamelCase : int = dict(self.forward_default_kwargs ) _UpperCamelCase : Any = kwargs.pop('''num_inference_steps''' , _snake_case ) _UpperCamelCase : Optional[int] = self.dummy_sample _UpperCamelCase : Tuple = 0.1 * sample _UpperCamelCase : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _UpperCamelCase : List[Any] = self.get_scheduler_config() _UpperCamelCase : Any = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals (must be after setting timesteps) _UpperCamelCase : List[Any] = dummy_past_residuals[:] if time_step is None: _UpperCamelCase : Dict = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) _UpperCamelCase : Any = scheduler_class.from_pretrained(_snake_case ) # copy over dummy past residuals new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residual (must be after setting timesteps) _UpperCamelCase : List[str] = dummy_past_residuals[:] _UpperCamelCase : List[Any] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : Optional[int] = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _UpperCamelCase : Any = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : List[str] = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self , **_snake_case ) -> Dict: _UpperCamelCase : List[str] = self.scheduler_classes[0] _UpperCamelCase : int = self.get_scheduler_config(**_snake_case ) _UpperCamelCase : Optional[int] = scheduler_class(**_snake_case ) _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : int = self.dummy_model() _UpperCamelCase : Optional[int] = self.dummy_sample_deter scheduler.set_timesteps(_snake_case ) for i, t in enumerate(scheduler.timesteps ): _UpperCamelCase : str = model(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scheduler.step(_snake_case , _snake_case , _snake_case ).prev_sample for i, t in enumerate(scheduler.timesteps ): _UpperCamelCase : Tuple = model(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scheduler.step(_snake_case , _snake_case , _snake_case ).prev_sample return sample def _lowercase ( self ) -> Optional[Any]: _UpperCamelCase : List[str] = dict(self.forward_default_kwargs ) _UpperCamelCase : Union[str, Any] = kwargs.pop('''num_inference_steps''' , _snake_case ) for scheduler_class in self.scheduler_classes: _UpperCamelCase : Union[str, Any] = self.get_scheduler_config() _UpperCamelCase : Dict = scheduler_class(**_snake_case ) _UpperCamelCase : Optional[Any] = self.dummy_sample _UpperCamelCase : Tuple = 0.1 * sample if num_inference_steps is not None and hasattr(_snake_case , '''set_timesteps''' ): scheduler.set_timesteps(_snake_case ) elif num_inference_steps is not None and not hasattr(_snake_case , '''set_timesteps''' ): _UpperCamelCase : str = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _UpperCamelCase : int = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] _UpperCamelCase : Optional[Any] = dummy_past_residuals[:] _UpperCamelCase : List[Any] = scheduler.timesteps[5] _UpperCamelCase : List[Any] = scheduler.timesteps[6] _UpperCamelCase : Optional[Any] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : Optional[int] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _UpperCamelCase : Optional[int] = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample _UpperCamelCase : Dict = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self ) -> int: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=_snake_case , time_step=_snake_case ) def _lowercase ( self ) -> Optional[Any]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=_snake_case , time_step=_snake_case ) def _lowercase ( self ) -> Optional[Any]: _UpperCamelCase : Optional[Any] = self.full_loop() _UpperCamelCase : List[Any] = torch.mean(torch.abs(_snake_case ) ) assert abs(result_mean.item() - 2540529 ) < 10
683
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers _UpperCAmelCase : Tuple = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) print("""NCCL version:""", torch.cuda.nccl.version()) except ImportError: print("""Torch version:""", None) try: import deepspeed print("""DeepSpeed version:""", deepspeed.__version__) except ImportError: print("""DeepSpeed version:""", None) try: import tensorflow as tf print("""TensorFlow version:""", tf.__version__) print("""TF GPUs available:""", bool(tf.config.list_physical_devices("""GPU"""))) print("""Number of TF GPUs available:""", len(tf.config.list_physical_devices("""GPU"""))) except ImportError: print("""TensorFlow version:""", None)
683
1
'''simple docstring''' def snake_case__ ( UpperCamelCase ) -> int: _UpperCamelCase : int = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def snake_case__ ( UpperCamelCase = 1_00 ) -> int: _UpperCamelCase : Any = 1 _UpperCamelCase : List[str] = 2 for i in range(2 ,max_n + 1 ): _UpperCamelCase : Tuple = pre_numerator _UpperCamelCase : Any = 2 * i // 3 if i % 3 == 0 else 1 _UpperCamelCase : int = cur_numerator _UpperCamelCase : List[Any] = e_cont * pre_numerator + temp return sum_digits(UpperCamelCase ) if __name__ == "__main__": print(f"""{solution() = }""")
683
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: return params[f'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase="attention" ) -> List[str]: _UpperCamelCase : Dict = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) _UpperCamelCase : int = k_tmp.reshape(k_tmp.shape[0] ,k_tmp.shape[1] * k_tmp.shape[2] ) _UpperCamelCase : str = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) _UpperCamelCase : Tuple = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] ,o_tmp.shape[2] ) _UpperCamelCase : Any = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) _UpperCamelCase : Optional[int] = q_tmp.reshape(q_tmp.shape[0] ,q_tmp.shape[1] * q_tmp.shape[2] ) _UpperCamelCase : Optional[Any] = np.ascontiguousarray(params[f'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) _UpperCamelCase : List[Any] = v_tmp.reshape(v_tmp.shape[0] ,v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=False ) -> List[str]: if split_mlp_wi: _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] _UpperCamelCase : Tuple = params[f'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] _UpperCamelCase : Optional[Any] = (wi_a, wi_a) else: _UpperCamelCase : str = params[f'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] _UpperCamelCase : int = params[f'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: return params[f'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def snake_case__ ( UpperCamelCase ,*, UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ) -> int: _UpperCamelCase : Any = traverse_util.flatten_dict(variables['''target'''] ) _UpperCamelCase : Optional[Any] = {'''/'''.join(UpperCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _UpperCamelCase : str = '''encoder/encoder/mlp/wi_0/kernel''' in old print('''Split MLP:''' ,UpperCamelCase ) _UpperCamelCase : Optional[int] = collections.OrderedDict() # Shared embeddings. _UpperCamelCase : str = old['''token_embedder/embedding'''] # Encoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Optional[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''attention''' ) _UpperCamelCase : Tuple = layer_norm _UpperCamelCase : int = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : Dict = v.T # Block i, layer 1 (MLP). _UpperCamelCase : Dict = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : int = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''encoder''' ,UpperCamelCase ) _UpperCamelCase : Union[str, Any] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Optional[Any] = wi[1].T else: _UpperCamelCase : List[Any] = wi.T _UpperCamelCase : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : Union[str, Any] = tax_relpos_bias_lookup( UpperCamelCase ,UpperCamelCase ,'''encoder''' ).T _UpperCamelCase : List[str] = old['''encoder/encoder_norm/scale'''] if not scalable_attention: _UpperCamelCase : List[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''encoder''' ).T _UpperCamelCase : Optional[Any] = tax_relpos_bias_lookup( UpperCamelCase ,0 ,'''decoder''' ).T if not is_encoder_only: # Decoder. for i in range(UpperCamelCase ): # Block i, layer 0 (Self Attention). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_self_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : List[Any] = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''self_attention''' ) _UpperCamelCase : int = layer_norm _UpperCamelCase : Union[str, Any] = k.T _UpperCamelCase : Optional[int] = o.T _UpperCamelCase : Dict = q.T _UpperCamelCase : Tuple = v.T # Block i, layer 1 (Cross Attention). _UpperCamelCase : str = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_cross_attention_layer_norm''' ) _UpperCamelCase, _UpperCamelCase, _UpperCamelCase, _UpperCamelCase : Dict = tax_attention_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''encoder_decoder_attention''' ) _UpperCamelCase : Dict = layer_norm _UpperCamelCase : Optional[int] = k.T _UpperCamelCase : int = o.T _UpperCamelCase : List[Any] = q.T _UpperCamelCase : str = v.T # Block i, layer 2 (MLP). _UpperCamelCase : Optional[int] = tax_layer_norm_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,'''pre_mlp_layer_norm''' ) _UpperCamelCase, _UpperCamelCase : List[Any] = tax_mlp_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ,UpperCamelCase ) _UpperCamelCase : List[str] = layer_norm if split_mlp_wi: _UpperCamelCase : Optional[Any] = wi[0].T _UpperCamelCase : Union[str, Any] = wi[1].T else: _UpperCamelCase : Dict = wi.T _UpperCamelCase : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer _UpperCamelCase : int = tax_relpos_bias_lookup(UpperCamelCase ,UpperCamelCase ,'''decoder''' ).T _UpperCamelCase : Optional[int] = old['''decoder/decoder_norm/scale'''] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _UpperCamelCase : str = old['''decoder/logits_dense/kernel'''].T return new def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Optional[int]: _UpperCamelCase : str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : str = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _UpperCamelCase : int = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''' ) _UpperCamelCase : Any = state_dict['''shared.weight'''] return state_dict def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Any: _UpperCamelCase : List[Any] = checkpoints.load_tax_checkpoint(UpperCamelCase ) _UpperCamelCase : str = convert_tax_to_pytorch( UpperCamelCase ,num_layers=config.num_layers ,is_encoder_only=UpperCamelCase ,scalable_attention=UpperCamelCase ) _UpperCamelCase : Optional[Any] = make_state_dict(UpperCamelCase ,UpperCamelCase ) model.load_state_dict(UpperCamelCase ,strict=UpperCamelCase ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = False ,UpperCamelCase = False ,) -> int: _UpperCamelCase : int = MTaConfig.from_json_file(UpperCamelCase ) print(f'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _UpperCamelCase : Optional[int] = UMTaEncoderModel(UpperCamelCase ) else: _UpperCamelCase : Optional[int] = UMTaForConditionalGeneration(UpperCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(UpperCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(UpperCamelCase ) print('''Done''' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
683
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Tuple = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } _UpperCAmelCase : Dict = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } _UpperCAmelCase : Tuple = """▁""" class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Optional[Any] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : int = AlbertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case=True , _snake_case=False , _snake_case="[CLS]" , _snake_case="[SEP]" , _snake_case="<unk>" , _snake_case="[SEP]" , _snake_case="<pad>" , _snake_case="[CLS]" , _snake_case="[MASK]" , **_snake_case , ) -> str: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Tuple = ( AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case , normalized=_snake_case ) if isinstance(_snake_case , _snake_case ) else mask_token ) super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , remove_space=_snake_case , keep_accents=_snake_case , bos_token=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , **_snake_case , ) _UpperCamelCase : List[str] = do_lower_case _UpperCamelCase : List[str] = remove_space _UpperCamelCase : int = keep_accents _UpperCamelCase : Optional[Any] = vocab_file _UpperCamelCase : Any = False if not self.vocab_file else True def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Tuple = [self.sep_token_id] _UpperCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Dict = [self.sep_token_id] _UpperCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_snake_case ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase : int = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_snake_case ): copyfile(self.vocab_file , _snake_case ) return (out_vocab_file,)
683
'''simple docstring''' from __future__ import annotations from functools import lru_cache from math import ceil _UpperCAmelCase : int = 100 _UpperCAmelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _UpperCAmelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_00 ) def snake_case__ ( UpperCamelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _UpperCamelCase : set[int] = set() _UpperCamelCase : int _UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def snake_case__ ( UpperCamelCase = 50_00 ) -> int | None: for number_to_partition in range(1 ,UpperCamelCase ): if len(partition(UpperCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
683
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class UpperCAmelCase ( a_ , unittest.TestCase ): """simple docstring""" A__ : Tuple = ShapEPipeline A__ : int = ['prompt'] A__ : Dict = ['prompt'] A__ : Union[str, Any] = [ 'num_images_per_prompt', 'num_inference_steps', 'generator', 'latents', 'guidance_scale', 'frame_size', 'output_type', 'return_dict', ] A__ : Any = False @property def _lowercase ( self ) -> str: return 32 @property def _lowercase ( self ) -> str: return 32 @property def _lowercase ( self ) -> Optional[Any]: return self.time_input_dim * 4 @property def _lowercase ( self ) -> Optional[Any]: return 8 @property def _lowercase ( self ) -> List[Any]: _UpperCamelCase : int = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def _lowercase ( self ) -> List[Any]: torch.manual_seed(0 ) _UpperCamelCase : List[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_snake_case ) @property def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = { '''num_attention_heads''': 2, '''attention_head_dim''': 16, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 32, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } _UpperCamelCase : Tuple = PriorTransformer(**_snake_case ) return model @property def _lowercase ( self ) -> int: torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = { '''param_shapes''': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 12, '''background''': ( 0.1, 0.1, 0.1, ), } _UpperCamelCase : List[Any] = ShapERenderer(**_snake_case ) return model def _lowercase ( self ) -> str: _UpperCamelCase : Dict = self.dummy_prior _UpperCamelCase : int = self.dummy_text_encoder _UpperCamelCase : Any = self.dummy_tokenizer _UpperCamelCase : Union[str, Any] = self.dummy_renderer _UpperCamelCase : Union[str, Any] = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=1024 , prediction_type='''sample''' , use_karras_sigmas=_snake_case , clip_sample=_snake_case , clip_sample_range=1.0 , ) _UpperCamelCase : Dict = { '''prior''': prior, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''renderer''': renderer, '''scheduler''': scheduler, } return components def _lowercase ( self , _snake_case , _snake_case=0 ) -> Optional[int]: if str(_snake_case ).startswith('''mps''' ): _UpperCamelCase : List[Any] = torch.manual_seed(_snake_case ) else: _UpperCamelCase : Any = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) _UpperCamelCase : int = { '''prompt''': '''horse''', '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 32, '''output_type''': '''np''', } return inputs def _lowercase ( self ) -> str: _UpperCamelCase : Tuple = '''cpu''' _UpperCamelCase : Optional[int] = self.get_dummy_components() _UpperCamelCase : int = self.pipeline_class(**_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = pipe(**self.get_dummy_inputs(_snake_case ) ) _UpperCamelCase : int = output.images[0] _UpperCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) _UpperCamelCase : Tuple = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self ) -> Union[str, Any]: # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : List[str] = torch_device == '''cpu''' _UpperCamelCase : Tuple = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_snake_case , relax_max_difference=_snake_case , ) def _lowercase ( self ) -> List[str]: _UpperCamelCase : int = self.get_dummy_components() _UpperCamelCase : Optional[Any] = self.pipeline_class(**_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = 1 _UpperCamelCase : str = 2 _UpperCamelCase : Optional[int] = self.get_dummy_inputs(_snake_case ) for key in inputs.keys(): if key in self.batch_params: _UpperCamelCase : Any = batch_size * [inputs[key]] _UpperCamelCase : List[Any] = pipe(**_snake_case , num_images_per_prompt=_snake_case )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: _UpperCamelCase : Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_np_out.npy''' ) _UpperCamelCase : Dict = ShapEPipeline.from_pretrained('''openai/shap-e''' ) _UpperCamelCase : Union[str, Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Union[str, Any] = torch.Generator(device=_snake_case ).manual_seed(0 ) _UpperCamelCase : int = pipe( '''a shark''' , generator=_snake_case , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type='''np''' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_snake_case , _snake_case )
683
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _UpperCAmelCase : Dict = """bart""" _UpperCAmelCase : List[str] = True @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> int: if LOAD_DENSE_INDEX: _UpperCamelCase : Optional[int] = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _UpperCamelCase : Optional[Any] = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _UpperCamelCase : Tuple = qar_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = (None, None) if MODEL_TYPE == "bart": _UpperCamelCase : Any = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _UpperCamelCase : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _UpperCamelCase : Dict = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _UpperCamelCase : Tuple = sas_model.eval() else: _UpperCamelCase, _UpperCamelCase : Optional[Any] = make_qa_sas_model( model_name='''t5-small''' ,from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' ,device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> List[Any]: if LOAD_DENSE_INDEX: _UpperCamelCase : str = faiss.StandardGpuResources() _UpperCamelCase : Optional[int] = datasets.load_dataset(path='''wiki_snippets''' ,name='''wiki40b_en_100_0''' )['''train'''] _UpperCamelCase : List[str] = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(wikiaab_passages.num_rows, 1_28) ,) _UpperCamelCase : Any = faiss.IndexFlatIP(1_28 ) _UpperCamelCase : str = faiss.index_cpu_to_gpu(UpperCamelCase ,1 ,UpperCamelCase ) wikiaab_gpu_index_flat.add(UpperCamelCase ) # TODO fix for larger GPU else: _UpperCamelCase, _UpperCamelCase : Optional[int] = (None, None) _UpperCamelCase : int = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=UpperCamelCase ) def snake_case__ ( ) -> Optional[int]: _UpperCamelCase : List[Any] = datasets.load_dataset('''eli5''' ,name='''LFQA_reddit''' ) _UpperCamelCase : Optional[int] = elia['''train_eli5'''] _UpperCamelCase : Any = np.memmap( '''eli5_questions_reps.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(elia_train.num_rows, 1_28) ) _UpperCamelCase : Optional[Any] = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(UpperCamelCase ) return (elia_train, eli5_train_q_index) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_indexes() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = load_models() _UpperCAmelCase , _UpperCAmelCase : int = load_train_data() def snake_case__ ( UpperCamelCase ,UpperCamelCase=10 ) -> Optional[Any]: _UpperCamelCase : Optional[int] = embed_questions_for_retrieval([question] ,UpperCamelCase ,UpperCamelCase ) _UpperCamelCase, _UpperCamelCase : Optional[Any] = eli5_train_q_index.search(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = [elia_train[int(UpperCamelCase )] for i in I[0]] return nn_examples def snake_case__ ( UpperCamelCase ,UpperCamelCase="wiki40b" ,UpperCamelCase="dense" ,UpperCamelCase=10 ) -> Optional[int]: if source == "none": _UpperCamelCase, _UpperCamelCase : Dict = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _UpperCamelCase, _UpperCamelCase : str = query_qa_dense_index( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) else: _UpperCamelCase, _UpperCamelCase : str = query_es_index( UpperCamelCase ,UpperCamelCase ,index_name='''english_wiki40b_snippets_100w''' ,n_results=UpperCamelCase ,) _UpperCamelCase : Optional[int] = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _UpperCamelCase : Optional[Any] = '''question: {} context: {}'''.format(UpperCamelCase ,UpperCamelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda UpperCamelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda UpperCamelCase : None), } ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase=64 ,UpperCamelCase=2_56 ,UpperCamelCase=False ,UpperCamelCase=2 ,UpperCamelCase=0.95 ,UpperCamelCase=0.8 ) -> Optional[Any]: with torch.no_grad(): _UpperCamelCase : Any = qa_sas_generate( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,num_answers=1 ,num_beams=UpperCamelCase ,min_len=UpperCamelCase ,max_len=UpperCamelCase ,do_sample=UpperCamelCase ,temp=UpperCamelCase ,top_p=UpperCamelCase ,top_k=UpperCamelCase ,max_input_length=10_24 ,device='''cuda:0''' ,)[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar _UpperCAmelCase : str = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _UpperCAmelCase : Tuple = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _UpperCAmelCase : Dict = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _UpperCAmelCase : List[str] = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _UpperCAmelCase : Optional[int] = st.sidebar.checkbox("""Demo options""") if demo_options: _UpperCAmelCase : List[str] = st.sidebar.selectbox( """""", action_list, index=3, ) _UpperCAmelCase : List[Any] = action_list.index(action_st) _UpperCAmelCase : Tuple = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) _UpperCAmelCase : Optional[Any] = show_type == """Show full text of passages""" else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : str = True _UpperCAmelCase : str = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: _UpperCAmelCase : Optional[Any] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) _UpperCAmelCase : Optional[Any] = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: _UpperCAmelCase : Dict = """wiki40b""" _UpperCAmelCase : str = """dense""" _UpperCAmelCase : List[str] = """beam""" _UpperCAmelCase : Dict = 2 _UpperCAmelCase : List[str] = 64 _UpperCAmelCase : List[Any] = 256 _UpperCAmelCase : Tuple = None _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : int = st.sidebar.checkbox("""Generation options""") if generate_options: _UpperCAmelCase : Union[str, Any] = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _UpperCAmelCase : str = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) _UpperCAmelCase : Dict = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) _UpperCAmelCase : List[Any] = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": _UpperCAmelCase : List[str] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[Any] = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _UpperCAmelCase : Optional[int] = None # start main text _UpperCAmelCase : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _UpperCAmelCase : int = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": _UpperCAmelCase : Any = st.text_input("""Enter your question here:""", """""") else: _UpperCAmelCase : Tuple = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": _UpperCAmelCase , _UpperCAmelCase : str = make_support(question, source=wiki_source, method="""dense""", n_results=10) _UpperCAmelCase , _UpperCAmelCase : List[Any] = make_support(question, source=wiki_source, method="""sparse""", n_results=10) _UpperCAmelCase : int = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _UpperCAmelCase : int = support_list[:10] _UpperCAmelCase : Tuple = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _UpperCAmelCase , _UpperCAmelCase : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): _UpperCAmelCase : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) _UpperCAmelCase : List[Any] = res[1].strip() if sec_titles == "": _UpperCAmelCase : Optional[int] = """[{}]({})""".format(res[0], wiki_url) else: _UpperCAmelCase : Optional[int] = sec_titles.split(""" & """) _UpperCAmelCase : Tuple = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: _UpperCAmelCase : Dict = find_nearest_training(question) _UpperCAmelCase : List[Any] = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) _UpperCAmelCase : List[Any] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) _UpperCAmelCase : List[Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
683
1
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : str = {"""configuration_mmbt""": ["""MMBTConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = ["""MMBTForClassification""", """MMBTModel""", """ModalEmbeddings"""] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
'''simple docstring''' from collections.abc import Iterable from typing import Any class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> Optional[int]: _UpperCamelCase : int = value _UpperCamelCase : Node | None = None # Added in order to delete a node easier _UpperCamelCase : Node | None = None _UpperCamelCase : Node | None = None def __repr__( self ) -> str: from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} , indent=1 ) class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case = None ) -> List[Any]: _UpperCamelCase : str = root def __str__( self ) -> str: return str(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if new_children is not None: # reset its kids _UpperCamelCase : Union[str, Any] = node.parent if node.parent is not None: # reset its parent if self.is_right(_snake_case ): # If it is the right children _UpperCamelCase : str = new_children else: _UpperCamelCase : Any = new_children else: _UpperCamelCase : Any = new_children def _lowercase ( self , _snake_case ) -> bool: if node.parent and node.parent.right: return node == node.parent.right return False def _lowercase ( self ) -> bool: return self.root is None def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : List[Any] = Node(_snake_case ) # create a new Node if self.empty(): # if Tree is empty _UpperCamelCase : Optional[Any] = new_node # set its root else: # Tree is not empty _UpperCamelCase : int = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: _UpperCamelCase : Union[str, Any] = new_node # We insert the new node in a leaf break else: _UpperCamelCase : Union[str, Any] = parent_node.left else: if parent_node.right is None: _UpperCamelCase : Any = new_node break else: _UpperCamelCase : str = parent_node.right _UpperCamelCase : Any = parent_node def _lowercase ( self , *_snake_case ) -> None: for value in values: self.__insert(_snake_case ) def _lowercase ( self , _snake_case ) -> Node | None: if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: _UpperCamelCase : List[str] = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: _UpperCamelCase : Optional[Any] = node.left if value < node.value else node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: if self.root is None: return None _UpperCamelCase : Dict = self.root if not self.empty(): while node.right is not None: _UpperCamelCase : Tuple = node.right return node def _lowercase ( self , _snake_case = None ) -> Node | None: if node is None: _UpperCamelCase : Optional[Any] = self.root if self.root is None: return None if not self.empty(): _UpperCamelCase : Optional[int] = self.root while node.left is not None: _UpperCamelCase : List[str] = node.left return node def _lowercase ( self , _snake_case ) -> None: _UpperCamelCase : str = self.search(_snake_case ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_snake_case , _snake_case ) elif node.left is None: # Has only right children self.__reassign_nodes(_snake_case , node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(_snake_case , node.left ) else: _UpperCamelCase : List[str] = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore _UpperCamelCase : int = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _lowercase ( self , _snake_case ) -> Iterable: if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _lowercase ( self , _snake_case=None ) -> Any: if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _lowercase ( self , _snake_case , _snake_case ) -> None: if node: self.inorder(_snake_case , node.left ) arr.append(node.value ) self.inorder(_snake_case , node.right ) def _lowercase ( self , _snake_case , _snake_case ) -> int: _UpperCamelCase : list[int] = [] self.inorder(_snake_case , _snake_case ) # append all values to list using inorder traversal return arr[k - 1] def snake_case__ ( UpperCamelCase ) -> list[Node]: _UpperCamelCase : int = [] if curr_node is not None: _UpperCamelCase : Any = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def snake_case__ ( ) -> None: _UpperCamelCase : Any = (8, 3, 6, 1, 10, 14, 13, 4, 7) _UpperCamelCase : Tuple = BinarySearchTree() for i in testlist: t.insert(UpperCamelCase ) # Prints all the elements of the list in order traversal print(UpperCamelCase ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' ,t.get_max().value ) # type: ignore print('''Min Value: ''' ,t.get_min().value ) # type: ignore for i in testlist: t.remove(UpperCamelCase ) print(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
683
1
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = None _UpperCamelCase : int = 20 _UpperCamelCase : Any = self._get_uniform_logits(batch_size=2 , length=_snake_case ) # tweak scores to not be uniform anymore _UpperCamelCase : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch _UpperCamelCase : Dict = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax _UpperCamelCase : Any = jax.nn.softmax(_snake_case , axis=-1 ) _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : List[str] = FlaxTemperatureLogitsWarper(temperature=1.3 ) _UpperCamelCase : List[str] = jax.nn.softmax(temp_dist_warper_sharper(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) _UpperCamelCase : str = jax.nn.softmax(temp_dist_warper_smoother(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _lowercase ( self ) -> Any: _UpperCamelCase : List[Any] = None _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : Any = 2 # create ramp distribution _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() _UpperCamelCase : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case _UpperCamelCase : Optional[int] = 5 _UpperCamelCase : str = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) _UpperCamelCase : Union[str, Any] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, length) ).copy() _UpperCamelCase : Optional[Any] = top_k_warp_safety_check(_snake_case , _snake_case , cur_len=_snake_case ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = None _UpperCamelCase : Any = 10 _UpperCamelCase : List[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) _UpperCamelCase : Tuple = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) _UpperCamelCase : List[str] = FlaxTopPLogitsWarper(0.8 ) _UpperCamelCase : Dict = np.exp(top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 _UpperCamelCase : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # check edge cases with negative and extreme logits _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme _UpperCamelCase : Tuple = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept _UpperCamelCase : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) _UpperCamelCase : Dict = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = 20 _UpperCamelCase : Optional[int] = 4 _UpperCamelCase : int = 0 _UpperCamelCase : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) # check that min length is applied at length 5 _UpperCamelCase : Any = ids_tensor((batch_size, 20) , vocab_size=20 ) _UpperCamelCase : int = 5 _UpperCamelCase : List[Any] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 _UpperCamelCase : Optional[int] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = 15 _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = 20 _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) # check that all scores are -inf except the bos_token_id score _UpperCamelCase : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) _UpperCamelCase : Optional[int] = 1 _UpperCamelCase : str = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : str = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 _UpperCamelCase : List[str] = 3 _UpperCamelCase : Tuple = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 20 _UpperCamelCase : Tuple = 4 _UpperCamelCase : Any = 0 _UpperCamelCase : str = 5 _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) # check that all scores are -inf except the eos_token_id when max_length is reached _UpperCamelCase : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) _UpperCamelCase : Dict = 4 _UpperCamelCase : Dict = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : int = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Any = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 4 _UpperCamelCase : Optional[Any] = 10 _UpperCamelCase : Dict = 15 _UpperCamelCase : Union[str, Any] = 2 _UpperCamelCase : Optional[Any] = 1 _UpperCamelCase : List[Any] = 15 # dummy input_ids and scores _UpperCamelCase : Optional[int] = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Any = input_ids.copy() _UpperCamelCase : int = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Tuple = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Optional[int] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : int = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : List[str] = 10 # no processor list _UpperCamelCase : Dict = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) # with processor list _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : Optional[Any] = processor(_snake_case , _snake_case , cur_len=_snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = 4 _UpperCamelCase : int = 10 _UpperCamelCase : List[Any] = 15 _UpperCamelCase : Dict = 2 _UpperCamelCase : Tuple = 1 _UpperCamelCase : Optional[int] = 15 # dummy input_ids and scores _UpperCamelCase : Tuple = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Optional[Any] = input_ids.copy() _UpperCamelCase : List[str] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[int] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : int = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : Union[str, Any] = 10 # no processor list def run_no_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : List[Any] = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) return scores # with processor list def run_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : List[str] = processor(_snake_case , _snake_case , cur_len=_snake_case ) return scores _UpperCamelCase : Dict = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jitted_run_no_processor_list(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = jitted_run_processor_list(_snake_case , _snake_case , _snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
683
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off _UpperCAmelCase : Dict = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 10563, 10786, 11420, 11709, 11907, 13163, 13697, 13700, 14808, 15306, 16410, 16791, 17992, 19203, 19510, 20724, 22305, 22935, 27007, 30109, 30420, 33409, 34949, 40283, 40493, 40549, 47282, 49146, 50257, 50359, 50360, 50361 ] _UpperCAmelCase : int = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362 ] class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = 'whisper' A__ : Tuple = ['past_key_values'] A__ : Optional[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , _snake_case=51865 , _snake_case=80 , _snake_case=6 , _snake_case=4 , _snake_case=6 , _snake_case=4 , _snake_case=1536 , _snake_case=1536 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=50257 , _snake_case=True , _snake_case=True , _snake_case="gelu" , _snake_case=256 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.0 , _snake_case=0.02 , _snake_case=False , _snake_case=1500 , _snake_case=448 , _snake_case=50256 , _snake_case=50256 , _snake_case=50256 , _snake_case=None , _snake_case=[220, 50256] , _snake_case=False , _snake_case=256 , _snake_case=False , _snake_case=0.05 , _snake_case=10 , _snake_case=2 , _snake_case=0.0 , _snake_case=10 , _snake_case=0 , _snake_case=7 , **_snake_case , ) -> Any: _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Union[str, Any] = num_mel_bins _UpperCamelCase : List[str] = d_model _UpperCamelCase : str = encoder_layers _UpperCamelCase : Optional[int] = encoder_attention_heads _UpperCamelCase : str = decoder_layers _UpperCamelCase : Tuple = decoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : Optional[int] = encoder_ffn_dim _UpperCamelCase : Any = dropout _UpperCamelCase : Optional[Any] = attention_dropout _UpperCamelCase : List[Any] = activation_dropout _UpperCamelCase : int = activation_function _UpperCamelCase : List[Any] = init_std _UpperCamelCase : Optional[int] = encoder_layerdrop _UpperCamelCase : str = decoder_layerdrop _UpperCamelCase : List[str] = use_cache _UpperCamelCase : Optional[Any] = encoder_layers _UpperCamelCase : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCamelCase : List[str] = max_source_positions _UpperCamelCase : Optional[Any] = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. _UpperCamelCase : str = classifier_proj_size _UpperCamelCase : List[str] = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCamelCase : int = apply_spec_augment _UpperCamelCase : str = mask_time_prob _UpperCamelCase : int = mask_time_length _UpperCamelCase : List[Any] = mask_time_min_masks _UpperCamelCase : List[str] = mask_feature_prob _UpperCamelCase : Optional[int] = mask_feature_length _UpperCamelCase : Union[str, Any] = mask_feature_min_masks _UpperCamelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , is_encoder_decoder=_snake_case , decoder_start_token_id=_snake_case , suppress_tokens=_snake_case , begin_suppress_tokens=_snake_case , **_snake_case , ) class UpperCAmelCase ( a_ ): """simple docstring""" @property def _lowercase ( self ) -> Mapping[str, Mapping[int, str]]: _UpperCamelCase : Dict = OrderedDict( [ ('''input_features''', {0: '''batch''', 1: '''feature_size''', 2: '''encoder_sequence'''}), ] ) if self.use_past: _UpperCamelCase : Tuple = {0: '''batch'''} else: _UpperCamelCase : Dict = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_snake_case , direction='''inputs''' ) return common_inputs def _lowercase ( self , _snake_case , _snake_case = -1 , _snake_case = -1 , _snake_case = False , _snake_case = None , _snake_case = 22050 , _snake_case = 5.0 , _snake_case = 220 , ) -> Mapping[str, Any]: _UpperCamelCase : Optional[int] = OrderedDict() _UpperCamelCase : Tuple = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_snake_case , framework=_snake_case , sampling_rate=_snake_case , time_duration=_snake_case , frequency=_snake_case , ) _UpperCamelCase : int = encoder_inputs['''input_features'''].shape[2] _UpperCamelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length _UpperCamelCase : str = super().generate_dummy_inputs( preprocessor.tokenizer , _snake_case , _snake_case , _snake_case , _snake_case ) _UpperCamelCase : Union[str, Any] = encoder_inputs.pop('''input_features''' ) _UpperCamelCase : Dict = decoder_inputs.pop('''decoder_input_ids''' ) if "past_key_values" in decoder_inputs: _UpperCamelCase : List[str] = decoder_inputs.pop('''past_key_values''' ) return dummy_inputs @property def _lowercase ( self ) -> float: return 1E-3
683
1
'''simple docstring''' from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image _UpperCAmelCase : int = ["""text""", """image""", """audio"""] def snake_case__ ( UpperCamelCase ) -> List[Any]: _UpperCamelCase : int = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((5_12, 5_12) ) ) elif input_type == "audio": inputs.append(torch.ones(30_00 ) ) elif isinstance(UpperCamelCase ,UpperCamelCase ): inputs.append(create_inputs(UpperCamelCase ) ) else: raise ValueError(f'''Invalid type requested: {input_type}''' ) return inputs def snake_case__ ( UpperCamelCase ) -> Any: _UpperCamelCase : int = [] for output in outputs: if isinstance(UpperCamelCase ,(str, AgentText) ): output_types.append('''text''' ) elif isinstance(UpperCamelCase ,(Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(UpperCamelCase ,(torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(f'''Invalid output: {output}''' ) return output_types @is_tool_test class UpperCAmelCase : """simple docstring""" def _lowercase ( self ) -> str: self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) _UpperCamelCase : List[Any] = self.tool.inputs for _input in inputs: if isinstance(_input , _snake_case ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) _UpperCamelCase : Optional[Any] = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def _lowercase ( self ) -> str: _UpperCamelCase : str = create_inputs(self.tool.inputs ) _UpperCamelCase : Optional[int] = self.tool(*_snake_case ) # There is a single output if len(self.tool.outputs ) == 1: _UpperCamelCase : Optional[Any] = [outputs] self.assertListEqual(output_types(_snake_case ) , self.tool.outputs ) def _lowercase ( self ) -> int: self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Union[str, Any] = create_inputs(self.tool.inputs ) _UpperCamelCase : Any = self.tool(*_snake_case ) if not isinstance(_snake_case , _snake_case ): _UpperCamelCase : Any = [outputs] self.assertEqual(len(_snake_case ) , len(self.tool.outputs ) ) for output, output_type in zip(_snake_case , self.tool.outputs ): _UpperCamelCase : Union[str, Any] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(_snake_case , _snake_case ) ) def _lowercase ( self ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = create_inputs(self.tool.inputs ) _UpperCamelCase : List[str] = [] for _input, input_type in zip(_snake_case , self.tool.inputs ): if isinstance(_snake_case , _snake_case ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error _UpperCamelCase : List[str] = self.tool(*_snake_case ) if not isinstance(_snake_case , _snake_case ): _UpperCamelCase : List[Any] = [outputs] self.assertEqual(len(_snake_case ) , len(self.tool.outputs ) )
683
'''simple docstring''' import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser( description=( """Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""roberta""", choices=["""roberta""", """gpt2"""]) parser.add_argument("""--model_name""", default="""roberta-large""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_roberta_048131723.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") _UpperCAmelCase : int = parser.parse_args() if args.model_type == "roberta": _UpperCAmelCase : Union[str, Any] = RobertaForMaskedLM.from_pretrained(args.model_name) _UpperCAmelCase : int = """roberta""" elif args.model_type == "gpt2": _UpperCAmelCase : Optional[int] = GPTaLMHeadModel.from_pretrained(args.model_name) _UpperCAmelCase : Optional[int] = """transformer""" _UpperCAmelCase : Tuple = model.state_dict() _UpperCAmelCase : int = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: _UpperCAmelCase : Optional[Any] = state_dict[f"""{prefix}.{param_name}"""] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: _UpperCAmelCase : Tuple = f"""{prefix}.embeddings.{w}.weight""" _UpperCAmelCase : Optional[Any] = state_dict[param_name] for w in ["weight", "bias"]: _UpperCAmelCase : Union[str, Any] = f"""{prefix}.embeddings.LayerNorm.{w}""" _UpperCAmelCase : str = state_dict[param_name] # Transformer Blocks # _UpperCAmelCase : Dict = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: _UpperCAmelCase : str = state_dict[ f"""{prefix}.h.{teacher_idx}.{layer}.{w}""" ] _UpperCAmelCase : Any = state_dict[f"""{prefix}.h.{teacher_idx}.attn.bias"""] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: _UpperCAmelCase : Optional[Any] = state_dict[ f"""{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}""" ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: _UpperCAmelCase : Dict = state_dict[f"""{layer}"""] if args.vocab_transform: for w in ["weight", "bias"]: _UpperCAmelCase : int = state_dict[f"""lm_head.dense.{w}"""] _UpperCAmelCase : int = state_dict[f"""lm_head.layer_norm.{w}"""] elif args.model_type == "gpt2": for w in ["weight", "bias"]: _UpperCAmelCase : List[str] = state_dict[f"""{prefix}.ln_f.{w}"""] _UpperCAmelCase : Any = state_dict["""lm_head.weight"""] print(f"""N layers selected for distillation: {std_idx}""") print(f"""Number of params transferred for distillation: {len(compressed_sd.keys())}""") print(f"""Save transferred checkpoint to {args.dump_checkpoint}.""") torch.save(compressed_sd, args.dump_checkpoint)
683
1
'''simple docstring''' import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _UpperCAmelCase : Any = logging.get_logger(__name__) class UpperCAmelCase ( a_ ): """simple docstring""" A__ : str = 'AutoTokenizer' A__ : int = ['tokenizer'] A__ : Tuple = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self , _snake_case , _snake_case=None ) -> Union[str, Any]: super().__init__(_snake_case ) _UpperCamelCase : Union[str, Any] = speaker_embeddings @classmethod def _lowercase ( cls , _snake_case , _snake_case="speaker_embeddings_path.json" , **_snake_case ) -> List[str]: if speaker_embeddings_dict_path is not None: _UpperCamelCase : Optional[int] = get_file_from_repo( _snake_case , _snake_case , subfolder=kwargs.pop('''subfolder''' , _snake_case ) , cache_dir=kwargs.pop('''cache_dir''' , _snake_case ) , force_download=kwargs.pop('''force_download''' , _snake_case ) , proxies=kwargs.pop('''proxies''' , _snake_case ) , resume_download=kwargs.pop('''resume_download''' , _snake_case ) , local_files_only=kwargs.pop('''local_files_only''' , _snake_case ) , use_auth_token=kwargs.pop('''use_auth_token''' , _snake_case ) , revision=kwargs.pop('''revision''' , _snake_case ) , ) if speaker_embeddings_path is None: logger.warning( F'''`{os.path.join(_snake_case , _snake_case )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _UpperCamelCase : Tuple = None else: with open(_snake_case ) as speaker_embeddings_json: _UpperCamelCase : Union[str, Any] = json.load(_snake_case ) else: _UpperCamelCase : Optional[int] = None _UpperCamelCase : int = AutoTokenizer.from_pretrained(_snake_case , **_snake_case ) return cls(tokenizer=_snake_case , speaker_embeddings=_snake_case ) def _lowercase ( self , _snake_case , _snake_case="speaker_embeddings_path.json" , _snake_case="speaker_embeddings" , _snake_case = False , **_snake_case , ) -> int: if self.speaker_embeddings is not None: os.makedirs(os.path.join(_snake_case , _snake_case , '''v2''' ) , exist_ok=_snake_case ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Any = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCamelCase : Optional[int] = self._load_voice_preset(_snake_case ) _UpperCamelCase : Optional[int] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['''repo_or_path'''] , _snake_case , F'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=_snake_case , ) _UpperCamelCase : Union[str, Any] = os.path.join(_snake_case , F'''{prompt_key}_{key}.npy''' ) _UpperCamelCase : Dict = tmp_dict with open(os.path.join(_snake_case , _snake_case ) , '''w''' ) as fp: json.dump(_snake_case , _snake_case ) super().save_pretrained(_snake_case , _snake_case , **_snake_case ) def _lowercase ( self , _snake_case = None , **_snake_case ) -> Any: _UpperCamelCase : int = self.speaker_embeddings[voice_preset] _UpperCamelCase : Tuple = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _UpperCamelCase : List[str] = get_file_from_repo( self.speaker_embeddings.get('''repo_or_path''' , '''/''' ) , voice_preset_paths[key] , subfolder=kwargs.pop('''subfolder''' , _snake_case ) , cache_dir=kwargs.pop('''cache_dir''' , _snake_case ) , force_download=kwargs.pop('''force_download''' , _snake_case ) , proxies=kwargs.pop('''proxies''' , _snake_case ) , resume_download=kwargs.pop('''resume_download''' , _snake_case ) , local_files_only=kwargs.pop('''local_files_only''' , _snake_case ) , use_auth_token=kwargs.pop('''use_auth_token''' , _snake_case ) , revision=kwargs.pop('''revision''' , _snake_case ) , ) if path is None: raise ValueError( F'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _UpperCamelCase : int = np.load(_snake_case ) return voice_preset_dict def _lowercase ( self , _snake_case = None ) -> Optional[int]: for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , _snake_case=None , _snake_case=None , _snake_case="pt" , _snake_case=256 , _snake_case=False , _snake_case=True , _snake_case=False , **_snake_case , ) -> List[Any]: if voice_preset is not None and not isinstance(_snake_case , _snake_case ): if ( isinstance(_snake_case , _snake_case ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCamelCase : Tuple = self._load_voice_preset(_snake_case ) else: if isinstance(_snake_case , _snake_case ) and not voice_preset.endswith('''.npz''' ): _UpperCamelCase : Optional[int] = voice_preset + '''.npz''' _UpperCamelCase : Any = np.load(_snake_case ) if voice_preset is not None: self._validate_voice_preset_dict(_snake_case , **_snake_case ) _UpperCamelCase : str = BatchFeature(data=_snake_case , tensor_type=_snake_case ) _UpperCamelCase : Union[str, Any] = self.tokenizer( _snake_case , return_tensors=_snake_case , padding='''max_length''' , max_length=_snake_case , return_attention_mask=_snake_case , return_token_type_ids=_snake_case , add_special_tokens=_snake_case , **_snake_case , ) if voice_preset is not None: _UpperCamelCase : str = voice_preset return encoded_text
683
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self , _snake_case , _snake_case ) -> Union[str, Any]: _UpperCamelCase : Optional[int] = jnp.ones((batch_size, length) ) / length return scores def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : int = None _UpperCamelCase : int = 20 _UpperCamelCase : Any = self._get_uniform_logits(batch_size=2 , length=_snake_case ) # tweak scores to not be uniform anymore _UpperCamelCase : Any = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch _UpperCamelCase : Dict = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax _UpperCamelCase : Any = jax.nn.softmax(_snake_case , axis=-1 ) _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : List[str] = FlaxTemperatureLogitsWarper(temperature=1.3 ) _UpperCamelCase : List[str] = jax.nn.softmax(temp_dist_warper_sharper(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) _UpperCamelCase : str = jax.nn.softmax(temp_dist_warper_smoother(_snake_case , scores.copy() , cur_len=_snake_case ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1E-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1E-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _lowercase ( self ) -> Any: _UpperCamelCase : List[Any] = None _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : Any = 2 # create ramp distribution _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() _UpperCamelCase : Union[str, Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case _UpperCamelCase : Optional[int] = 5 _UpperCamelCase : str = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) _UpperCamelCase : Union[str, Any] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, length) ).copy() _UpperCamelCase : Optional[Any] = top_k_warp_safety_check(_snake_case , _snake_case , cur_len=_snake_case ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _lowercase ( self ) -> Optional[int]: _UpperCamelCase : Any = None _UpperCamelCase : Any = 10 _UpperCamelCase : List[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) _UpperCamelCase : Tuple = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) _UpperCamelCase : List[str] = FlaxTopPLogitsWarper(0.8 ) _UpperCamelCase : Dict = np.exp(top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 _UpperCamelCase : Optional[int] = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # check edge cases with negative and extreme logits _UpperCamelCase : Optional[int] = np.broadcast_to(np.arange(_snake_case )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme _UpperCamelCase : Tuple = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept _UpperCamelCase : Tuple = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) _UpperCamelCase : Dict = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _lowercase ( self ) -> Dict: _UpperCamelCase : List[Any] = 20 _UpperCamelCase : Optional[int] = 4 _UpperCamelCase : int = 0 _UpperCamelCase : Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) # check that min length is applied at length 5 _UpperCamelCase : Any = ids_tensor((batch_size, 20) , vocab_size=20 ) _UpperCamelCase : int = 5 _UpperCamelCase : List[Any] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 _UpperCamelCase : Optional[int] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = 15 _UpperCamelCase : Optional[int] = min_dist_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> List[Any]: _UpperCamelCase : Optional[int] = 20 _UpperCamelCase : Union[str, Any] = 4 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) # check that all scores are -inf except the bos_token_id score _UpperCamelCase : Union[str, Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) _UpperCamelCase : Optional[int] = 1 _UpperCamelCase : str = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : str = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 _UpperCamelCase : List[str] = 3 _UpperCamelCase : Tuple = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 20 _UpperCamelCase : Tuple = 4 _UpperCamelCase : Any = 0 _UpperCamelCase : str = 5 _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) # check that all scores are -inf except the eos_token_id when max_length is reached _UpperCamelCase : Optional[Any] = ids_tensor((batch_size, 4) , vocab_size=20 ) _UpperCamelCase : Dict = 4 _UpperCamelCase : Dict = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : int = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Any = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[Any] = logits_processor(_snake_case , _snake_case , cur_len=_snake_case ) self.assertFalse(jnp.isinf(_snake_case ).any() ) def _lowercase ( self ) -> str: _UpperCamelCase : Dict = 4 _UpperCamelCase : Optional[Any] = 10 _UpperCamelCase : Dict = 15 _UpperCamelCase : Union[str, Any] = 2 _UpperCamelCase : Optional[Any] = 1 _UpperCamelCase : List[Any] = 15 # dummy input_ids and scores _UpperCamelCase : Optional[int] = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Any = input_ids.copy() _UpperCamelCase : int = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : List[str] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Tuple = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : Optional[int] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Tuple = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : int = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Tuple = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : List[str] = 10 # no processor list _UpperCamelCase : Dict = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Optional[int] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) # with processor list _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : Optional[Any] = processor(_snake_case , _snake_case , cur_len=_snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _lowercase ( self ) -> Tuple: _UpperCamelCase : Tuple = 4 _UpperCamelCase : int = 10 _UpperCamelCase : List[Any] = 15 _UpperCamelCase : Dict = 2 _UpperCamelCase : Tuple = 1 _UpperCamelCase : Optional[int] = 15 # dummy input_ids and scores _UpperCamelCase : Tuple = ids_tensor((batch_size, sequence_length) , _snake_case ) _UpperCamelCase : Optional[Any] = input_ids.copy() _UpperCamelCase : List[str] = self._get_uniform_logits(_snake_case , _snake_case ) _UpperCamelCase : Optional[int] = scores.copy() # instantiate all dist processors _UpperCamelCase : Optional[int] = FlaxTemperatureLogitsWarper(temperature=0.5 ) _UpperCamelCase : Dict = FlaxTopKLogitsWarper(3 ) _UpperCamelCase : int = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors _UpperCamelCase : Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_snake_case ) _UpperCamelCase : Optional[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_snake_case ) _UpperCamelCase : Any = FlaxForcedEOSTokenLogitsProcessor(max_length=_snake_case , eos_token_id=_snake_case ) _UpperCamelCase : Union[str, Any] = 10 # no processor list def run_no_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : List[Any] = temp_dist_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Tuple = top_k_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = top_p_warp(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = min_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : Union[str, Any] = bos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) _UpperCamelCase : str = eos_dist_proc(_snake_case , _snake_case , cur_len=_snake_case ) return scores # with processor list def run_processor_list(_snake_case , _snake_case , _snake_case ): _UpperCamelCase : Optional[Any] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) _UpperCamelCase : List[str] = processor(_snake_case , _snake_case , cur_len=_snake_case ) return scores _UpperCamelCase : Dict = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jax.jit(_snake_case ) _UpperCamelCase : Optional[int] = jitted_run_no_processor_list(_snake_case , _snake_case , _snake_case ) _UpperCamelCase : Any = jitted_run_processor_list(_snake_case , _snake_case , _snake_case ) # scores should be equal self.assertTrue(jnp.allclose(_snake_case , _snake_case , atol=1E-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
683
1
'''simple docstring''' import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py _UpperCAmelCase : Union[str, Any] = """\ @INPROCEEDINGS{Papineni02bleu:a, author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu}, title = {BLEU: a Method for Automatic Evaluation of Machine Translation}, booktitle = {}, year = {2002}, pages = {311--318} } @inproceedings{lin-och-2004-orange, title = \"{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation\", author = \"Lin, Chin-Yew and Och, Franz Josef\", booktitle = \"{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics\", month = \"aug 23{--}aug 27\", year = \"2004\", address = \"Geneva, Switzerland\", publisher = \"COLING\", url = \"https://www.aclweb.org/anthology/C04-1072\", pages = \"501--507\", } """ _UpperCAmelCase : Optional[Any] = """\ BLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another. Quality is considered to be the correspondence between a machine's output and that of a human: \"the closer a machine translation is to a professional human translation, the better it is\" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and remains one of the most popular automated and inexpensive metrics. Scores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations. Those scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality. Intelligibility or grammatical correctness are not taken into account[citation needed]. BLEU's output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1 representing more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the reference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional reference translations will increase the BLEU score. """ _UpperCAmelCase : int = """ Computes BLEU score of translated segments against one or more references. Args: predictions: list of translations to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. max_order: Maximum n-gram order to use when computing BLEU score. smooth: Whether or not to apply Lin et al. 2004 smoothing. Returns: 'bleu': bleu score, 'precisions': geometric mean of n-gram precisions, 'brevity_penalty': brevity penalty, 'length_ratio': ratio of lengths, 'translation_length': translation_length, 'reference_length': reference_length Examples: >>> predictions = [ ... [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample ... [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample ... ] >>> references = [ ... [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references) ... [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference) ... ] >>> bleu = datasets.load_metric(\"bleu\") >>> results = bleu.compute(predictions=predictions, references=references) >>> print(results[\"bleu\"]) 1.0 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): """simple docstring""" def _lowercase ( self ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ), '''references''': datasets.Sequence( datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=['''https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def _lowercase ( self , _snake_case , _snake_case , _snake_case=4 , _snake_case=False ) -> List[Any]: _UpperCamelCase : Optional[int] = compute_bleu( reference_corpus=_snake_case , translation_corpus=_snake_case , max_order=_snake_case , smooth=_snake_case ) ((_UpperCamelCase), (_UpperCamelCase), (_UpperCamelCase), (_UpperCamelCase), (_UpperCamelCase), (_UpperCamelCase)) : Optional[int] = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
683
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _UpperCAmelCase : Optional[int] = pytest.mark.integration @pytest.mark.parametrize('''path''' ,['''paws''', '''csv'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Dict: inspect_dataset(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' ,['''accuracy'''] ) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> int: inspect_metric(UpperCamelCase ,UpperCamelCase ) _UpperCamelCase : List[str] = path + '''.py''' assert script_name in os.listdir(UpperCamelCase ) assert "__pycache__" not in os.listdir(UpperCamelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> int: _UpperCamelCase : List[str] = get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[str]: with pytest.raises(UpperCamelCase ): get_dataset_config_info(UpperCamelCase ,config_name=UpperCamelCase ) @pytest.mark.parametrize( '''path, expected''' ,[ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : int = get_dataset_config_names(UpperCamelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' ,[ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Dict: _UpperCamelCase : Dict = get_dataset_infos(UpperCamelCase ) assert list(infos.keys() ) == expected_configs _UpperCamelCase : Dict = expected_configs[0] assert expected_config in infos _UpperCamelCase : Any = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' ,[ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> Union[str, Any]: _UpperCamelCase : List[Any] = get_dataset_infos(UpperCamelCase ) assert expected_config in infos _UpperCamelCase : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' ,[ ('''paws''', None, ValueError), ] ,) def snake_case__ ( UpperCamelCase ,UpperCamelCase ,UpperCamelCase ) -> List[Any]: with pytest.raises(UpperCamelCase ): get_dataset_split_names(UpperCamelCase ,config_name=UpperCamelCase )
683
1
'''simple docstring''' from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent _UpperCAmelCase : Union[str, Any] = {"""UserAgent""": UserAgent().random} def snake_case__ ( UpperCamelCase ) -> dict: _UpperCamelCase : Dict = script.contents[0] _UpperCamelCase : Union[str, Any] = json.loads(data[data.find('''{"config"''' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class UpperCAmelCase : """simple docstring""" def __init__( self , _snake_case ) -> Any: _UpperCamelCase : List[Any] = F'''https://www.instagram.com/{username}/''' _UpperCamelCase : List[Any] = self.get_json() def _lowercase ( self ) -> dict: _UpperCamelCase : Optional[Any] = requests.get(self.url , headers=_snake_case ).text _UpperCamelCase : Dict = BeautifulSoup(_snake_case , '''html.parser''' ).find_all('''script''' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self ) -> str: return F'''{self.__class__.__name__}(\'{self.username}\')''' def __str__( self ) -> str: return F'''{self.fullname} ({self.username}) is {self.biography}''' @property def _lowercase ( self ) -> str: return self.user_data["username"] @property def _lowercase ( self ) -> str: return self.user_data["full_name"] @property def _lowercase ( self ) -> str: return self.user_data["biography"] @property def _lowercase ( self ) -> str: return self.user_data["business_email"] @property def _lowercase ( self ) -> str: return self.user_data["external_url"] @property def _lowercase ( self ) -> int: return self.user_data["edge_followed_by"]["count"] @property def _lowercase ( self ) -> int: return self.user_data["edge_follow"]["count"] @property def _lowercase ( self ) -> int: return self.user_data["edge_owner_to_timeline_media"]["count"] @property def _lowercase ( self ) -> str: return self.user_data["profile_pic_url_hd"] @property def _lowercase ( self ) -> bool: return self.user_data["is_verified"] @property def _lowercase ( self ) -> bool: return self.user_data["is_private"] def snake_case__ ( UpperCamelCase = "github" ) -> None: import os if os.environ.get('''CI''' ): return # test failing on GitHub Actions _UpperCamelCase : Tuple = InstagramUser(UpperCamelCase ) assert instagram_user.user_data assert isinstance(instagram_user.user_data ,UpperCamelCase ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 1_50 assert instagram_user.number_of_followers > 12_00_00 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('''https://instagram.''' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() _UpperCAmelCase : int = InstagramUser("""github""") print(instagram_user) print(f"""{instagram_user.number_of_posts = }""") print(f"""{instagram_user.number_of_followers = }""") print(f"""{instagram_user.number_of_followings = }""") print(f"""{instagram_user.email = }""") print(f"""{instagram_user.website = }""") print(f"""{instagram_user.profile_picture_url = }""") print(f"""{instagram_user.is_verified = }""") print(f"""{instagram_user.is_private = }""")
683
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ) -> Dict: torch.manual_seed(0 ) _UpperCamelCase : Any = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return model @property def _lowercase ( self ) -> Tuple: torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , cross_attention_dim=10 , ) return model @property def _lowercase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) _UpperCamelCase : Optional[int] = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') , up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') , ) _UpperCamelCase : int = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return vqvae, unet @slow def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Tuple = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _UpperCamelCase : int = DDPMScheduler() _UpperCamelCase : Optional[int] = AudioDiffusionPipeline(vqvae=_snake_case , unet=self.dummy_unet , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case , steps=4 ) _UpperCamelCase : Union[str, Any] = output.audios[0] _UpperCamelCase : Union[str, Any] = output.images[0] _UpperCamelCase : str = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : int = pipe(generator=_snake_case , steps=4 , return_dict=_snake_case ) _UpperCamelCase : int = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _UpperCamelCase : List[str] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : List[str] = np.frombuffer(image_from_tuple.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : int = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : str = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _UpperCamelCase : Dict = DDIMScheduler() _UpperCamelCase : str = self.dummy_vqvae_and_unet _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : List[Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : Optional[Any] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _UpperCamelCase : Optional[Any] = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Tuple = pipe(raw_audio=_snake_case , generator=_snake_case , start_step=5 , steps=10 ) _UpperCamelCase : List[str] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _UpperCamelCase : Any = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Tuple = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _UpperCamelCase : Any = self.dummy_unet_condition _UpperCamelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_snake_case , mel=_snake_case , scheduler=_snake_case ) _UpperCamelCase : Union[str, Any] = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) np.random.seed(0 ) _UpperCamelCase : int = torch.rand((1, 1, 10) ) _UpperCamelCase : Optional[Any] = pipe(generator=_snake_case , encoding=_snake_case ) _UpperCamelCase : Dict = output.images[0] _UpperCamelCase : Dict = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Any = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: _UpperCamelCase : Optional[int] = torch_device _UpperCamelCase : int = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) _UpperCamelCase : str = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) _UpperCamelCase : Tuple = torch.Generator(device=_snake_case ).manual_seed(42 ) _UpperCamelCase : Optional[int] = pipe(generator=_snake_case ) _UpperCamelCase : List[Any] = output.audios[0] _UpperCamelCase : List[Any] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _UpperCamelCase : Union[str, Any] = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] _UpperCamelCase : Union[str, Any] = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
683
1
'''simple docstring''' _UpperCAmelCase : str = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ _UpperCAmelCase : Any = [{"""type""": """code""", """content""": INSTALL_CONTENT}] _UpperCAmelCase : Optional[int] = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
683
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : Tuple = { """configuration_longt5""": ["""LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongT5Config""", """LongT5OnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongT5EncoderModel""", """LongT5ForConditionalGeneration""", """LongT5Model""", """LongT5PreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """FlaxLongT5ForConditionalGeneration""", """FlaxLongT5Model""", """FlaxLongT5PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
683
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : List[Any] = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all BART models at https://huggingface.co/models?filter=bart _UpperCAmelCase : Any = { """vocab_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/vocab.json""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/vocab.json""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json""", }, """merges_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/merges.txt""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/merges.txt""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt""", }, """tokenizer_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json""", }, } _UpperCAmelCase : List[str] = { """facebook/bart-base""": 1024, """facebook/bart-large""": 1024, """facebook/bart-large-mnli""": 1024, """facebook/bart-large-cnn""": 1024, """facebook/bart-large-xsum""": 1024, """yjernite/bart_eli5""": 1024, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : Dict = VOCAB_FILES_NAMES A__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP A__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = ['input_ids', 'attention_mask'] A__ : Dict = BartTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case="replace" , _snake_case="<s>" , _snake_case="</s>" , _snake_case="</s>" , _snake_case="<s>" , _snake_case="<unk>" , _snake_case="<pad>" , _snake_case="<mask>" , _snake_case=False , _snake_case=True , **_snake_case , ) -> Optional[int]: super().__init__( _snake_case , _snake_case , tokenizer_file=_snake_case , errors=_snake_case , bos_token=_snake_case , eos_token=_snake_case , sep_token=_snake_case , cls_token=_snake_case , unk_token=_snake_case , pad_token=_snake_case , mask_token=_snake_case , add_prefix_space=_snake_case , trim_offsets=_snake_case , **_snake_case , ) _UpperCamelCase : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , _snake_case ) != add_prefix_space: _UpperCamelCase : List[Any] = getattr(_snake_case , pre_tok_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = add_prefix_space _UpperCamelCase : int = pre_tok_class(**_snake_case ) _UpperCamelCase : Optional[Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` _UpperCamelCase : Optional[int] = '''post_processor''' _UpperCamelCase : Union[str, Any] = getattr(self.backend_tokenizer , _snake_case , _snake_case ) if tokenizer_component_instance: _UpperCamelCase : Dict = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _UpperCamelCase : Tuple = tuple(state['''sep'''] ) if "cls" in state: _UpperCamelCase : Any = tuple(state['''cls'''] ) _UpperCamelCase : Optional[Any] = False if state.get('''add_prefix_space''' , _snake_case ) != add_prefix_space: _UpperCamelCase : Tuple = add_prefix_space _UpperCamelCase : str = True if state.get('''trim_offsets''' , _snake_case ) != trim_offsets: _UpperCamelCase : List[Any] = trim_offsets _UpperCamelCase : Optional[int] = True if changes_to_apply: _UpperCamelCase : Dict = getattr(_snake_case , state.pop('''type''' ) ) _UpperCamelCase : Optional[Any] = component_class(**_snake_case ) setattr(self.backend_tokenizer , _snake_case , _snake_case ) @property def _lowercase ( self ) -> str: if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def _lowercase ( self , _snake_case ) -> Any: _UpperCamelCase : int = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case ) if isinstance(_snake_case , _snake_case ) else value _UpperCamelCase : Tuple = value def _lowercase ( self , *_snake_case , **_snake_case ) -> BatchEncoding: _UpperCamelCase : Optional[int] = kwargs.get('''is_split_into_words''' , _snake_case ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*_snake_case , **_snake_case ) def _lowercase ( self , *_snake_case , **_snake_case ) -> BatchEncoding: _UpperCamelCase : int = kwargs.get('''is_split_into_words''' , _snake_case ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*_snake_case , **_snake_case ) def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : str = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case ) def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[Any]: _UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
683
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } _UpperCAmelCase : Any = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class UpperCAmelCase ( a_ ): """simple docstring""" A__ : List[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : Tuple = DistilBertTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ) -> int: super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _UpperCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , _snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _snake_case ) != tokenize_chinese_chars ): _UpperCamelCase : int = getattr(_snake_case , normalizer_state.pop('''type''' ) ) _UpperCamelCase : Optional[int] = do_lower_case _UpperCamelCase : Dict = strip_accents _UpperCamelCase : List[Any] = tokenize_chinese_chars _UpperCamelCase : Tuple = normalizer_class(**_snake_case ) _UpperCamelCase : Dict = do_lower_case def _lowercase ( self , _snake_case , _snake_case=None ) -> Optional[int]: _UpperCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , _snake_case , _snake_case = None ) -> List[int]: _UpperCamelCase : Union[str, Any] = [self.sep_token_id] _UpperCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , _snake_case , _snake_case = None ) -> Tuple[str]: _UpperCamelCase : Optional[Any] = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
683
1