code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations __UpperCAmelCase = [] def UpperCamelCase ( snake_case__ : list[list[int]] , snake_case__ : int , snake_case__ : int ) -> bool: for i in range(len(snake_case__ ) ): if board[row][i] == 1: return False for i in range(len(snake_case__ ) ): if board[i][column] == 1: return False for i, j in zip(range(snake_case__ , -1 , -1 ) , range(snake_case__ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(snake_case__ , -1 , -1 ) , range(snake_case__ , len(snake_case__ ) ) ): if board[i][j] == 1: return False return True def UpperCamelCase ( snake_case__ : list[list[int]] , snake_case__ : int ) -> bool: if row >= len(snake_case__ ): solution.append(snake_case__ ) printboard(snake_case__ ) print() return True for i in range(len(snake_case__ ) ): if is_safe(snake_case__ , snake_case__ , snake_case__ ): UpperCamelCase : int = 1 solve(snake_case__ , row + 1 ) UpperCamelCase : Dict = 0 return False def UpperCamelCase ( snake_case__ : list[list[int]] ) -> None: for i in range(len(snake_case__ ) ): for j in range(len(snake_case__ ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) __UpperCAmelCase = 8 __UpperCAmelCase = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('''The total no. of solutions are :''', len(solution))
40
from ...configuration_utils import PretrainedConfig UpperCAmelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "tapas" def __init__( self : List[Any] , A__ : str=3_05_22 , A__ : Tuple=7_68 , A__ : List[Any]=12 , A__ : Optional[Any]=12 , A__ : Union[str, Any]=30_72 , A__ : Dict="gelu" , A__ : List[Any]=0.1 , A__ : str=0.1 , A__ : List[Any]=10_24 , A__ : Optional[int]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , A__ : Union[str, Any]=0.02 , A__ : Tuple=1E-12 , A__ : Tuple=0 , A__ : Any=10.0 , A__ : List[str]=0 , A__ : List[str]=1.0 , A__ : Optional[Any]=None , A__ : Tuple=1.0 , A__ : Union[str, Any]=False , A__ : Any=None , A__ : Union[str, Any]=1.0 , A__ : int=1.0 , A__ : str=False , A__ : int=False , A__ : Optional[Any]="ratio" , A__ : str=None , A__ : int=None , A__ : Dict=64 , A__ : int=32 , A__ : Optional[Any]=False , A__ : List[str]=True , A__ : List[Any]=False , A__ : str=False , A__ : Any=True , A__ : Tuple=False , A__ : str=None , A__ : str=None , **A__ : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=A__ , **A__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) snake_case_ : int = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[int] = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : List[Any] = type_vocab_sizes snake_case_ : str = initializer_range snake_case_ : Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters snake_case_ : Optional[int] = positive_label_weight snake_case_ : Dict = num_aggregation_labels snake_case_ : List[str] = aggregation_loss_weight snake_case_ : str = use_answer_as_supervision snake_case_ : int = answer_loss_importance snake_case_ : Any = use_normalized_answer_loss snake_case_ : int = huber_loss_delta snake_case_ : List[Any] = temperature snake_case_ : str = aggregation_temperature snake_case_ : List[str] = use_gumbel_for_cells snake_case_ : List[str] = use_gumbel_for_aggregation snake_case_ : Dict = average_approximation_function snake_case_ : List[str] = cell_selection_preference snake_case_ : Dict = answer_loss_cutoff snake_case_ : List[str] = max_num_rows snake_case_ : Union[str, Any] = max_num_columns snake_case_ : str = average_logits_per_cell snake_case_ : Union[str, Any] = select_one_column snake_case_ : Dict = allow_empty_column_selection snake_case_ : List[Any] = init_cell_selection_weights_to_zero snake_case_ : str = reset_position_index_per_cell snake_case_ : List[Any] = disable_per_token_loss # Aggregation hyperparameters snake_case_ : List[str] = aggregation_labels snake_case_ : Union[str, Any] = no_aggregation_label_index if isinstance(self.aggregation_labels , A__ ): snake_case_ : Optional[int] = {int(A__ ): v for k, v in aggregation_labels.items()}
666
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = XGLMTokenizer SCREAMING_SNAKE_CASE : int = XGLMTokenizerFast SCREAMING_SNAKE_CASE : List[str] = True SCREAMING_SNAKE_CASE : Tuple = True def SCREAMING_SNAKE_CASE ( self : Tuple ): super().setUp() # We have a SentencePiece fixture for testing __lowercase = XGLMTokenizer(lowercase__ ,keep_accents=lowercase__ ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = '''<pad>''' __lowercase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase__ ) ,lowercase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase__ ) ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'''<s>''' ) self.assertEqual(vocab_keys[1] ,'''<pad>''' ) self.assertEqual(len(lowercase__ ) ,1_0_0_8 ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,1_0_0_8 ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = XGLMTokenizer(lowercase__ ,keep_accents=lowercase__ ) __lowercase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowercase__ ,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase__ ) ,[value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] ,) __lowercase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowercase__ ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] ,) __lowercase = tokenizer.convert_tokens_to_ids(lowercase__ ) self.assertListEqual( lowercase__ ,[ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] ,) __lowercase = tokenizer.convert_ids_to_tokens(lowercase__ ) self.assertListEqual( lowercase__ ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] ,) @cached_property def SCREAMING_SNAKE_CASE ( self : int ): return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowercase__ ,f.name ) __lowercase = XGLMTokenizer(f.name ,keep_accents=lowercase__ ) __lowercase = pickle.dumps(lowercase__ ) pickle.loads(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = '''I was born in 92000, and this is falsé.''' __lowercase = tokenizer.tokenize(lowercase__ ) __lowercase = rust_tokenizer.tokenize(lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) __lowercase = tokenizer.encode(lowercase__ ,add_special_tokens=lowercase__ ) __lowercase = rust_tokenizer.encode(lowercase__ ,add_special_tokens=lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase__ ) __lowercase = rust_tokenizer.encode(lowercase__ ) self.assertListEqual(lowercase__ ,lowercase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = '''Hello World!''' __lowercase = [2, 3_1_2_2_7, 4_4_4_7, 3_5] self.assertListEqual(lowercase__ ,self.big_tokenizer.encode(lowercase__ ) ) @slow def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off __lowercase = [2, 1_0_1_8, 6_7, 1_1, 1_9_8_8, 2_6_1_7, 5_6_3_1, 2_7_8, 1_1, 3_4_0_7, 4_8, 7_1_6_3_0, 2_8_0_8_5, 4, 3_2_3_4, 1_5_7, 1_3, 6, 5, 6, 4, 3_5_2_6, 7_6_8, 1_5, 6_5_9, 5_7, 2_9_8, 3_9_8_3, 8_6_4, 1_2_9, 2_1, 6, 5, 1_3_6_7_5, 3_7_7, 6_5_2, 7_5_8_0, 1_0_3_4_1, 1_5_5, 2_8_1_7, 4_2_2, 1_6_6_6, 7, 1_6_7_4, 5_3, 1_1_3, 2_0_2_2_7_7, 1_7_8_9_2, 3_3, 6_0, 8_7, 4, 3_2_3_4, 1_5_7, 6_1, 2_6_6_7, 5_2_3_7_6, 1_9, 8_8, 2_3, 7_3_5] # fmt: on self.assertListEqual(lowercase__ ,self.big_tokenizer.encode(lowercase__ ) ) @slow def SCREAMING_SNAKE_CASE ( self : List[Any] ): # fmt: off __lowercase = { '''input_ids''': [[2, 1_0_8_8_2_5, 1_1_6_3, 1_5, 8_8_0_1_0, 4_7_3, 1_5_8_9_8, 1_5_7, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 2_3_8_0_2_1, 1_1_6_3, 5_3, 1_3_6_7_2, 1_8_5_7, 3_1_2, 8, 5_3_2_8_3, 1_8_2_3_9_6, 8, 1_8_5_6_6, 1_6, 3_6_7_3_3, 4_1_0_1, 8, 2_3_0, 2_4_4_0_1_7, 1_2_2_5_5_3, 7, 1_5, 1_3_2_5_9_7, 4, 2_9_3, 1_2_5_1_1, 7_6_1_0, 4, 3_4_1_4, 1_3_2_5_9_7, 9, 4, 3_2_3_6_1, 3_6_2, 4, 7_3_4, 2_8_5_1_2, 3_2_5_6_9, 1_8, 4, 3_2_3_6_1, 2_6_0_9_6, 1_4_9_8_2, 7_3, 1_8_7_1_5, 2_1_4_3_3, 2_3_5_2_6_1, 1_5, 4_9_2, 1_2_4_2_7, 1_6, 5_3, 1_8_7_1_5, 2_1_4_3_3, 6_5_4_5_4, 1_5, 2_3_6_5_9, 5_6_3, 1_6, 2_7_8, 5_9_7, 2_8_4_3, 5_9_5, 7_9_3_1, 1_8_2_3_9_6, 6_4_1_8_6, 2_2, 8_8_6, 5_9_5, 1_3_2_9_8_1, 5_3, 2_5_5_4_0, 3_4_4_9, 4_3_9_8_2, 3_9_9_0_1, 5_9_5_1, 8_7_8, 3_3_0, 4, 2_7_6_9_4, 8_0_2_6_9, 3_1_2, 5_3, 6_5_1_7, 1_1_7_8_0, 6_1_1, 2_0_4_0_8, 5], [2, 6, 1_3_2_5_9_7, 6_7, 4_2_8_9_7, 3_3, 5_9_2, 8, 1_6_3_7_2_9, 2_5_5_4_0, 3_6_1, 1_3_6_9_9_7, 1_0_9_5_1_4, 1_7_3_2_3_0, 7, 5_0_1, 6_0, 1_0_2_9_1_3, 1_9_6, 5_6_3_1, 2_3_5, 6_3_2_4_3, 4_7_3, 6, 2_3_1_7_5_7, 7_4, 5_2_7_7, 7_9_0_5, 5_3, 3_0_9_5, 3_7_3_1_7, 2_2, 4_5_4, 1_8_3_8_7_4, 5], [2, 2_6_8, 3_1_2_9_8, 4_6_5_3_0, 6, 1_3_2_9_3_5, 4_3_8_3_1, 7, 5_9_7, 3_2, 2_4, 3_6_8_8, 9_8_6_5, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowercase__ ,model_name='''facebook/xglm-564M''' ,padding=lowercase__ ,)
41
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : str , A__ : str ) -> Optional[int]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : int , A__ : Optional[int] , A__ : Dict ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Any , A__ : List[str] , A__ : str ) -> Optional[int]: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : List[Any] , A__ : List[str] , A__ : Optional[int] ) -> List[str]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class snake_case__ ( _UpperCamelCase ): @require_beam def UpperCAmelCase__ ( self : str ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Dict = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[int] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : str ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam snake_case_ : Tuple = beam.io.parquetio.WriteToParquet snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[Any] = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: snake_case_ : int = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[Any] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Tuple = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[str] = NestedBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) snake_case_ : int = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
666
0
'''simple docstring''' import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class UpperCAmelCase : '''simple docstring''' def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Any: '''simple docstring''' return None class UpperCAmelCase : '''simple docstring''' def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: '''simple docstring''' return None class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def UpperCamelCase( self ) -> str: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(SCREAMING_SNAKE_CASE_ , 'tf' , 12 , **SCREAMING_SNAKE_CASE_ ) @require_torch @slow def UpperCamelCase( self ) -> List[str]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(SCREAMING_SNAKE_CASE_ , 'pt' , 12 , **SCREAMING_SNAKE_CASE_ ) @require_torch @slow def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' from transformers import BertModel lowerCamelCase_ = ['[UNK]', '[SEP]', '[CLS]', '[PAD]', '[MASK]', 'some', 'other', 'words'] with NamedTemporaryFile(mode='w+t' ) as vocab_file: vocab_file.write('\n'.join(SCREAMING_SNAKE_CASE_ ) ) vocab_file.flush() lowerCamelCase_ = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: lowerCamelCase_ = BertModel(BertConfig(vocab_size=len(SCREAMING_SNAKE_CASE_ ) ) ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) self._test_export(SCREAMING_SNAKE_CASE_ , 'pt' , 12 , SCREAMING_SNAKE_CASE_ ) @require_tf @slow def UpperCamelCase( self ) -> int: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowerCamelCase_ = self._test_export(SCREAMING_SNAKE_CASE_ , 'tf' , 12 , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = quantize(Path(SCREAMING_SNAKE_CASE_ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(SCREAMING_SNAKE_CASE_ ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) @require_torch @slow def UpperCamelCase( self ) -> str: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowerCamelCase_ = self._test_export(SCREAMING_SNAKE_CASE_ , 'pt' , 12 , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = quantize(SCREAMING_SNAKE_CASE_ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(SCREAMING_SNAKE_CASE_ ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: '''simple docstring''' try: # Compute path with TemporaryDirectory() as tempdir: lowerCamelCase_ = Path(SCREAMING_SNAKE_CASE_ ).joinpath('model.onnx' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) return path except Exception as e: self.fail(SCREAMING_SNAKE_CASE_ ) @require_torch @require_tokenizers @slow def UpperCamelCase( self ) -> Tuple: '''simple docstring''' from transformers import BertModel lowerCamelCase_ = BertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'pt' ) @require_tf @require_tokenizers @slow def UpperCamelCase( self ) -> Optional[Any]: '''simple docstring''' from transformers import TFBertModel lowerCamelCase_ = TFBertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowerCamelCase_ = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'tf' ) def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Dict: '''simple docstring''' lowerCamelCase_ = FeatureExtractionPipeline(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = ['input_ids', 'token_type_ids', 'attention_mask', 'output_0', 'output_1'] lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = infer_shapes(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Assert all variables are present self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , len(SCREAMING_SNAKE_CASE_ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , SCREAMING_SNAKE_CASE_ ) self.assertSequenceEqual(variable_names[3:] , SCREAMING_SNAKE_CASE_ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: 'batch', 1: 'sequence'} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['output_0'] , {0: 'batch', 1: 'sequence'} ) self.assertDictEqual(shapes['output_1'] , {0: 'batch'} ) def UpperCamelCase( self ) -> List[str]: '''simple docstring''' lowerCamelCase_ = ['input_ids', 'attention_mask', 'token_type_ids'] lowerCamelCase_ = {'input_ids': [1, 2, 3, 4], 'attention_mask': [0, 0, 0, 0], 'token_type_ids': [1, 1, 1, 1]} lowerCamelCase_ ,lowerCamelCase_ = ensure_valid_input(FuncContiguousArgs() , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(SCREAMING_SNAKE_CASE_ ) , set(SCREAMING_SNAKE_CASE_ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(SCREAMING_SNAKE_CASE_ , (tokens['input_ids'], tokens['token_type_ids'], tokens['attention_mask']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) lowerCamelCase_ ,lowerCamelCase_ = ensure_valid_input(FuncNonContiguousArgs() , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 1 ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['input_ids'] ) self.assertEqual(ordered_input_names[0] , 'input_ids' ) def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = generate_identified_filename(Path('/home/something/my_fake_model.onnx' ) , '-test' ) self.assertEqual('/home/something/my_fake_model-test.onnx' , generated.as_posix() )
42
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple ): class snake_case__ : def __init__( self : Any , A__ : List[str] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = metric_id class snake_case__ : _SCREAMING_SNAKE_CASE : List[str] = [MetricMock(_UpperCamelCase ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: int , lowerCAmelCase_: List[Any] , lowerCAmelCase_: Any , lowerCAmelCase_: List[str] ): if "tmp_path" in args: snake_case_ : List[Any] = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowerCAmelCase_ , match="https://huggingface.co/docs/evaluate" ): func(*lowerCAmelCase_ )
666
0
from __future__ import annotations from typing import Any class _a : def __init__( self: int , UpperCamelCase_: int ) -> None: """simple docstring""" lowercase__ = num_of_nodes lowercase__ = [] lowercase__ = {} def lowerCamelCase_ ( self: Tuple , UpperCamelCase_: int , UpperCamelCase_: int , UpperCamelCase_: int ) -> None: """simple docstring""" self.m_edges.append([u_node, v_node, weight] ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase_: int ) -> int: """simple docstring""" if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def lowerCamelCase_ ( self: str , UpperCamelCase_: int ) -> None: """simple docstring""" if self.m_component[u_node] != u_node: for k in self.m_component: lowercase__ = self.find_component(UpperCamelCase_ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase_: list[int] , UpperCamelCase_: int , UpperCamelCase_: int ) -> None: """simple docstring""" if component_size[u_node] <= component_size[v_node]: lowercase__ = v_node component_size[v_node] += component_size[u_node] self.set_component(UpperCamelCase_ ) elif component_size[u_node] >= component_size[v_node]: lowercase__ = self.find_component(UpperCamelCase_ ) component_size[u_node] += component_size[v_node] self.set_component(UpperCamelCase_ ) def lowerCamelCase_ ( self: Tuple ) -> None: """simple docstring""" lowercase__ = [] lowercase__ = 0 lowercase__ = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) lowercase__ = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: lowercase__ , lowercase__ , lowercase__ = edge lowercase__ = self.m_component[u] lowercase__ = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): lowercase__ = [u, v, w] for edge in minimum_weight_edge: if isinstance(UpperCamelCase_ , UpperCamelCase_ ): lowercase__ , lowercase__ , lowercase__ = edge lowercase__ = self.m_component[u] lowercase__ = self.m_component[v] if u_component != v_component: mst_weight += w self.union(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) print(f'Added edge [{u} - {v}]\nAdded weight: {w}\n' ) num_of_components -= 1 lowercase__ = [-1] * self.m_num_of_nodes print(f'The total weight of the minimal spanning tree is: {mst_weight}' ) def _a ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
43
from __future__ import annotations import bisect def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Any = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: snake_case_ : Tuple = mid + 1 else: snake_case_ : Dict = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Optional[Any] = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: snake_case_ : Optional[Any] = mid + 1 else: snake_case_ : Tuple = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_left(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_right(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Dict = 0 snake_case_ : Tuple = len(lowerCAmelCase_ ) - 1 while left <= right: snake_case_ : int = left + (right - left) // 2 snake_case_ : Optional[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: snake_case_ : Optional[Any] = midpoint - 1 else: snake_case_ : Optional[int] = midpoint + 1 return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Optional[int] = bisect.bisect_left(lowerCAmelCase_ , lowerCAmelCase_ ) if index != len(lowerCAmelCase_ ) and sorted_collection[index] == item: return index return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int , lowerCAmelCase_: int ): if right < left: return None snake_case_ : List[Any] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , midpoint - 1 ) else: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , midpoint + 1 , lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = input("Enter numbers separated by comma:\n").strip() UpperCAmelCase = sorted(int(item) for item in user_input.split(",")) UpperCAmelCase = int(input("Enter a single number to be found in the list:\n")) UpperCAmelCase = binary_search(collection, target) if result is None: print(F"{target} was not found in {collection}.") else: print(F"{target} was found at position {result} in {collection}.")
666
0
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def A_ ( _lowerCAmelCase : Union[str, Any] ): """simple docstring""" _lowerCamelCase : str = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def A_ ( _lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : List[str] = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: _lowerCamelCase : Tuple = s_dict.pop(_lowerCAmelCase ) elif "subsample" in key: _lowerCamelCase : Tuple = s_dict.pop(_lowerCAmelCase ) def A_ ( _lowerCAmelCase : Union[str, Any] ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Optional[int] = emb.weight.shape _lowerCamelCase : Optional[Any] = nn.Linear(_lowerCAmelCase , _lowerCAmelCase , bias=_lowerCAmelCase ) _lowerCamelCase : Optional[Any] = emb.weight.data return lin_layer def A_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] ): """simple docstring""" _lowerCamelCase : Union[str, Any] = torch.load(_lowerCAmelCase , map_location="cpu" ) _lowerCamelCase : Tuple = mam_aaa["args"] _lowerCamelCase : str = mam_aaa["model"] _lowerCamelCase : List[str] = state_dict["decoder.output_projection.weight"] remove_ignore_keys_(_lowerCAmelCase ) rename_keys(_lowerCAmelCase ) _lowerCamelCase : int = state_dict["decoder.embed_tokens.weight"].shape[0] _lowerCamelCase : List[Any] = args.share_decoder_input_output_embed _lowerCamelCase : Union[str, Any] = [int(_lowerCAmelCase ) for i in args.conv_kernel_sizes.split("," )] _lowerCamelCase : Dict = SpeechaTextConfig( vocab_size=_lowerCAmelCase , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="relu" , num_conv_layers=len(_lowerCAmelCase ) , conv_channels=args.conv_channels , conv_kernel_sizes=_lowerCAmelCase , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=_lowerCAmelCase , num_beams=5 , max_length=200 , use_cache=_lowerCAmelCase , decoder_start_token_id=2 , early_stopping=_lowerCAmelCase , ) _lowerCamelCase : str = SpeechaTextForConditionalGeneration(_lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : Dict = model.model.load_state_dict(_lowerCAmelCase , strict=_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0 and not set(_lowerCAmelCase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," F' but all the following weights are missing {missing}' ) if tie_embeds: _lowerCamelCase : Optional[Any] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: _lowerCamelCase : Tuple = lm_head_weights model.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase_ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') UpperCAmelCase_ : int = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
44
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , A__ : VQModel , A__ : UNetaDModel , A__ : DDIMScheduler ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , A__ : int = 1 , A__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A__ : float = 0.0 , A__ : int = 50 , A__ : Optional[str] = "pil" , A__ : bool = True , **A__ : Optional[Any] , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' snake_case_ : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) snake_case_ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler snake_case_ : Any = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature snake_case_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case_ : List[Any] = {} if accepts_eta: snake_case_ : int = eta for t in self.progress_bar(self.scheduler.timesteps ): snake_case_ : Union[str, Any] = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual snake_case_ : Dict = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case_ : Union[str, Any] = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE snake_case_ : int = self.vqvae.decode(A__ ).sample snake_case_ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ : Optional[int] = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
666
0
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : List[str] = """vision-encoder-decoder""" _snake_case : Optional[Any] = True def __init__( self :List[Any] , **lowerCamelCase__ :List[str] ): super().__init__(**lowerCamelCase__ ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"""A configuraton of type {self.model_type} cannot be instantiated because """ f"""not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}""" ) UpperCamelCase__ :List[Any] = kwargs.pop("""encoder""" ) UpperCamelCase__ :Any = encoder_config.pop("""model_type""" ) UpperCamelCase__ :Optional[int] = kwargs.pop("""decoder""" ) UpperCamelCase__ :Optional[Any] = decoder_config.pop("""model_type""" ) UpperCamelCase__ :Optional[Any] = AutoConfig.for_model(lowerCamelCase__ , **lowerCamelCase__ ) UpperCamelCase__ :List[Any] = AutoConfig.for_model(lowerCamelCase__ , **lowerCamelCase__ ) UpperCamelCase__ :Dict = True @classmethod def __a ( cls :Union[str, Any] , lowerCamelCase__ :PretrainedConfig , lowerCamelCase__ :PretrainedConfig , **lowerCamelCase__ :List[Any] ): logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) UpperCamelCase__ :Dict = True UpperCamelCase__ :Optional[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **lowerCamelCase__ ) def __a ( self :Dict ): UpperCamelCase__ :Any = copy.deepcopy(self.__dict__ ) UpperCamelCase__ :str = self.encoder.to_dict() UpperCamelCase__ :Any = self.decoder.to_dict() UpperCamelCase__ :List[Any] = self.__class__.model_type return output class lowerCAmelCase_ ( lowercase ): """simple docstring""" _snake_case : List[Any] = version.parse("""1.11""" ) @property def __a ( self :int ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __a ( self :Union[str, Any] ): return 1e-4 @property def __a ( self :List[Any] ): return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class lowerCAmelCase_ ( lowercase ): """simple docstring""" @property def __a ( self :Optional[Any] ): UpperCamelCase__ :str = OrderedDict() UpperCamelCase__ :Union[str, Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCamelCase__ :List[str] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} UpperCamelCase__ :str = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def __a ( self :int , lowerCamelCase__ :"PreTrainedTokenizerBase" , lowerCamelCase__ :int = -1 , lowerCamelCase__ :int = -1 , lowerCamelCase__ :bool = False , lowerCamelCase__ :Optional["TensorType"] = None , ): import torch UpperCamelCase__ :Dict = OrderedDict() UpperCamelCase__ :Tuple = super().generate_dummy_inputs( lowerCamelCase__ , batch_size=lowerCamelCase__ , seq_length=lowerCamelCase__ , is_pair=lowerCamelCase__ , framework=lowerCamelCase__ ) UpperCamelCase__ , UpperCamelCase__ :str = dummy_input["""input_ids"""].shape UpperCamelCase__ :str = (batch, encoder_sequence, self._config.encoder_hidden_size) UpperCamelCase__ :Optional[Any] = dummy_input.pop("""input_ids""" ) UpperCamelCase__ :List[Any] = dummy_input.pop("""attention_mask""" ) UpperCamelCase__ :str = torch.zeros(lowerCamelCase__ ) return common_inputs class lowerCAmelCase_ ( lowercase ): """simple docstring""" @property def __a ( self :List[str] ): pass def __a ( self :List[str] , lowerCamelCase__ :PretrainedConfig ): return VisionEncoderDecoderEncoderOnnxConfig(lowerCamelCase__ ) def __a ( self :Dict , lowerCamelCase__ :PretrainedConfig , lowerCamelCase__ :PretrainedConfig , lowerCamelCase__ :str = "default" ): UpperCamelCase__ :Tuple = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(lowerCamelCase__ , lowerCamelCase__ )
45
from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) snake_case_ : List[str] = precision snake_case_ : Union[str, Any] = ceil(precision / 1_4 ) snake_case_ : List[str] = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : str = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : str = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): snake_case_ : Tuple = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCAmelCase = 5_0 print(F"The first {n} digits of pi is: {pi(n)}")
666
0
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase : List[Any] = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
46
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int = 1_0_0_0 ): snake_case_ ,snake_case_ : List[str] = 1, 1 snake_case_ : List[str] = 2 while True: snake_case_ : Tuple = 0 snake_case_ : Union[str, Any] = fa + fa snake_case_ ,snake_case_ : str = fa, f index += 1 for _ in str(lowerCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
666
0
import os from pathlib import Path def UpperCAmelCase__ ( ): from torch.utils.cpp_extension import load __a : Tuple = Path(lowerCamelCase_ ).resolve().parent.parent.parent / 'kernels' / 'deformable_detr' __a : Tuple = [ root / filename for filename in [ 'vision.cpp', os.path.join('cpu' , 'ms_deform_attn_cpu.cpp' ), os.path.join('cuda' , 'ms_deform_attn_cuda.cu' ), ] ] load( 'MultiScaleDeformableAttention' , lowerCamelCase_ , with_cuda=lowerCamelCase_ , extra_include_paths=[str(lowerCamelCase_ )] , extra_cflags=['-DWITH_CUDA=1'] , extra_cuda_cflags=[ '-DCUDA_HAS_FP16=1', '-D__CUDA_NO_HALF_OPERATORS__', '-D__CUDA_NO_HALF_CONVERSIONS__', '-D__CUDA_NO_HALF2_OPERATORS__', ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
47
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int | float] , lowerCAmelCase_: int , lowerCAmelCase_: int ): if len(lowerCAmelCase_ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(lowerCAmelCase_ ) or left < -len(lowerCAmelCase_ ) or right >= len(lowerCAmelCase_ ) or right < -len(lowerCAmelCase_ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case_ : List[Any] = (left + right) >> 1 # the middle snake_case_ : Dict = find_max(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # find max in range[left, mid] snake_case_ : int = find_max(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
666
0
'''simple docstring''' def A ( UpperCamelCase_ : list[int] ) -> list[int]: '''simple docstring''' lowerCAmelCase__ = len(UpperCamelCase_ ) for i in range(UpperCamelCase_ ): for j in range(i + 1 , UpperCamelCase_ ): if numbers[j] < numbers[i]: lowerCAmelCase__ ,lowerCAmelCase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": UpperCAmelCase__ : Dict = input("Enter numbers separated by a comma:\n").strip() UpperCAmelCase__ : str = [int(item) for item in user_input.split(",")] print(exchange_sort(unsorted))
48
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } UpperCAmelCase = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE : List[str] = RobertaTokenizer def __init__( self : Optional[int] , A__ : List[Any]=None , A__ : Optional[int]=None , A__ : List[str]=None , A__ : Dict="replace" , A__ : List[str]="<s>" , A__ : Optional[Any]="</s>" , A__ : List[str]="</s>" , A__ : List[Any]="<s>" , A__ : int="<unk>" , A__ : int="<pad>" , A__ : List[Any]="<mask>" , A__ : Any=False , A__ : Optional[int]=True , **A__ : Union[str, Any] , ) -> int: '''simple docstring''' super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : List[Any] = getattr(A__ , pre_tok_state.pop("type" ) ) snake_case_ : Any = add_prefix_space snake_case_ : List[Any] = pre_tok_class(**A__ ) snake_case_ : Optional[int] = add_prefix_space snake_case_ : List[str] = "post_processor" snake_case_ : Tuple = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case_ : str = tuple(state["sep"] ) if "cls" in state: snake_case_ : Tuple = tuple(state["cls"] ) snake_case_ : Tuple = False if state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : Optional[Any] = add_prefix_space snake_case_ : str = True if state.get("trim_offsets" , A__ ) != trim_offsets: snake_case_ : Optional[int] = trim_offsets snake_case_ : List[Any] = True if changes_to_apply: snake_case_ : int = getattr(A__ , state.pop("type" ) ) snake_case_ : List[Any] = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCAmelCase__ ( self : Optional[Any] ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCAmelCase__ ( self : Tuple , A__ : Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case_ : Any = value def UpperCAmelCase__ ( self : int , *A__ : Optional[Any] , **A__ : int ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[Any] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , *A__ : Any , **A__ : List[Any] ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[int] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCAmelCase__ ( self : int , A__ : List[str] , A__ : Union[str, Any]=None ) -> Any: '''simple docstring''' snake_case_ : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Dict , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : str = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
666
0
"""simple docstring""" from itertools import product def lowercase__ ( snake_case_ :int , snake_case_ :int ): __UpperCAmelCase = sides_number __UpperCAmelCase = max_face_number * dice_number __UpperCAmelCase = [0] * (max_total + 1) __UpperCAmelCase = 1 __UpperCAmelCase = range(snake_case_ , max_face_number + 1 ) for dice_numbers in product(snake_case_ , repeat=snake_case_ ): __UpperCAmelCase = sum(snake_case_ ) totals_frequencies[total] += 1 return totals_frequencies def lowercase__ ( ): __UpperCAmelCase = total_frequency_distribution( sides_number=4 , dice_number=9 ) __UpperCAmelCase = total_frequency_distribution( sides_number=6 , dice_number=6 ) __UpperCAmelCase = 0 __UpperCAmelCase = 9 __UpperCAmelCase = 4 * 9 __UpperCAmelCase = 6 for peter_total in range(snake_case_ , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) __UpperCAmelCase = (4**9) * (6**6) __UpperCAmelCase = peter_wins_count / total_games_number __UpperCAmelCase = round(snake_case_ , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(f"""{solution() = }""")
49
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) snake_case_ : Any = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } snake_case_ : List[str] = model(A__ )["last_hidden_state"] snake_case_ : str = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , A__ ) # compare the actual values for a slice. snake_case_ : List[str] = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
666
0
'''simple docstring''' import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ): lowerCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(_lowerCAmelCase ) lowerCamelCase__ = -1 lowerCamelCase__ = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(_lowerCAmelCase ) lowerCamelCase__ = model.generate(_lowerCAmelCase ,max_new_tokens=10 ,do_sample=_lowerCAmelCase ) lowerCamelCase__ = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: lowerCamelCase__ = TextStreamer(_lowerCAmelCase ) model.generate(_lowerCAmelCase ,max_new_tokens=10 ,do_sample=_lowerCAmelCase ,streamer=_lowerCAmelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCamelCase__ = cs.out[:-1] self.assertEqual(_lowerCAmelCase ,_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(_lowerCAmelCase ) lowerCamelCase__ = -1 lowerCamelCase__ = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(_lowerCAmelCase ) lowerCamelCase__ = model.generate(_lowerCAmelCase ,max_new_tokens=10 ,do_sample=_lowerCAmelCase ) lowerCamelCase__ = tokenizer.decode(greedy_ids[0] ) lowerCamelCase__ = TextIteratorStreamer(_lowerCAmelCase ) lowerCamelCase__ = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} lowerCamelCase__ = Thread(target=model.generate ,kwargs=_lowerCAmelCase ) thread.start() lowerCamelCase__ = """""" for new_text in streamer: streamer_text += new_text self.assertEqual(_lowerCAmelCase ,_lowerCAmelCase ) def UpperCamelCase_ ( self ): lowerCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(_lowerCAmelCase ) lowerCamelCase__ = -1 lowerCamelCase__ = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(_lowerCAmelCase ) lowerCamelCase__ = model.generate(_lowerCAmelCase ,max_new_tokens=10 ,do_sample=_lowerCAmelCase ) lowerCamelCase__ = greedy_ids[:, input_ids.shape[1] :] lowerCamelCase__ = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: lowerCamelCase__ = TextStreamer(_lowerCAmelCase ,skip_prompt=_lowerCAmelCase ) model.generate(_lowerCAmelCase ,max_new_tokens=10 ,do_sample=_lowerCAmelCase ,streamer=_lowerCAmelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCamelCase__ = cs.out[:-1] self.assertEqual(_lowerCAmelCase ,_lowerCAmelCase ) def UpperCamelCase_ ( self ): # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them lowerCamelCase__ = AutoTokenizer.from_pretrained("""distilgpt2""" ) lowerCamelCase__ = AutoModelForCausalLM.from_pretrained("""distilgpt2""" ).to(_lowerCAmelCase ) lowerCamelCase__ = -1 lowerCamelCase__ = torch.ones((1, 5) ,device=_lowerCAmelCase ).long() * model.config.bos_token_id with CaptureStdout() as cs: lowerCamelCase__ = TextStreamer(_lowerCAmelCase ,skip_special_tokens=_lowerCAmelCase ) model.generate(_lowerCAmelCase ,max_new_tokens=1 ,do_sample=_lowerCAmelCase ,streamer=_lowerCAmelCase ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token lowerCamelCase__ = cs.out[:-1] # Remove the final "\n" lowerCamelCase__ = tokenizer(_lowerCAmelCase ,return_tensors="""pt""" ) self.assertEqual(streamer_text_tokenized.input_ids.shape ,(1, 1) ) def UpperCamelCase_ ( self ): lowerCamelCase__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) lowerCamelCase__ = AutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ).to(_lowerCAmelCase ) lowerCamelCase__ = -1 lowerCamelCase__ = ids_tensor((1, 5) ,vocab_size=model.config.vocab_size ).to(_lowerCAmelCase ) lowerCamelCase__ = TextIteratorStreamer(_lowerCAmelCase ,timeout=0.001 ) lowerCamelCase__ = {"""input_ids""": input_ids, """max_new_tokens""": 10, """do_sample""": False, """streamer""": streamer} lowerCamelCase__ = Thread(target=model.generate ,kwargs=_lowerCAmelCase ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(_lowerCAmelCase ): lowerCamelCase__ = """""" for new_text in streamer: streamer_text += new_text
50
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = ["pixel_values"] def __init__( self : Union[str, Any] , A__ : bool = True , A__ : Dict[str, int] = None , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = True , **A__ : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : str = size if size is not None else {"shortest_edge": 2_24} snake_case_ : Union[str, Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : str = size snake_case_ : Optional[Any] = resample snake_case_ : Any = do_center_crop snake_case_ : Any = crop_size snake_case_ : str = do_rescale snake_case_ : Optional[Any] = rescale_factor snake_case_ : int = do_normalize snake_case_ : Optional[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ : List[str] = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ : int = do_convert_rgb def UpperCAmelCase__ ( self : Optional[int] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : str = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : np.ndarray , A__ : Union[int, float] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> str: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Any , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Any , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[Any] , A__ : ImageInput , A__ : bool = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : int = None , A__ : bool = None , A__ : float = None , A__ : bool = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A__ : Optional[Any] , ) -> PIL.Image.Image: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Union[str, Any] = size if size is not None else self.size snake_case_ : Any = get_size_dict(A__ , param_name="size" , default_to_square=A__ ) snake_case_ : Optional[int] = resample if resample is not None else self.resample snake_case_ : int = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" , default_to_square=A__ ) snake_case_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : List[Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ : List[Any] = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ : Dict = [convert_to_rgb(A__ ) for image in images] # All transformations expect numpy arrays. snake_case_ : Dict = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : Tuple = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : int = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : List[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Tuple = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ )
666
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a__ : Dict = logging.get_logger(__name__) a__ : Tuple = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' _lowerCamelCase ="yolos" def __init__( self : Any , a__ : List[Any]=768 , a__ : str=12 , a__ : Tuple=12 , a__ : Union[str, Any]=3072 , a__ : int="gelu" , a__ : Tuple=0.0 , a__ : Dict=0.0 , a__ : Dict=0.02 , a__ : Tuple=1e-1_2 , a__ : str=[512, 864] , a__ : List[Any]=16 , a__ : Dict=3 , a__ : str=True , a__ : Union[str, Any]=100 , a__ : Tuple=True , a__ : List[str]=False , a__ : Optional[Any]=1 , a__ : Dict=5 , a__ : int=2 , a__ : str=5 , a__ : Tuple=2 , a__ : Tuple=0.1 , **a__ : Optional[int] , ): super().__init__(**a__ ) UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = qkv_bias UpperCAmelCase = num_detection_tokens UpperCAmelCase = use_mid_position_embeddings UpperCAmelCase = auxiliary_loss # Hungarian matcher UpperCAmelCase = class_cost UpperCAmelCase = bbox_cost UpperCAmelCase = giou_cost # Loss coefficients UpperCAmelCase = bbox_loss_coefficient UpperCAmelCase = giou_loss_coefficient UpperCAmelCase = eos_coefficient class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' _lowerCamelCase =version.parse("1.11" ) @property def __snake_case ( self : str ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def __snake_case ( self : List[str] ): return 1e-4 @property def __snake_case ( self : Union[str, Any] ): return 12
51
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): snake_case_ ,snake_case_ : Dict = position snake_case_ : int = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] snake_case_ : Union[str, Any] = [] for position in positions: snake_case_ ,snake_case_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(lowerCAmelCase_ ) return permissible_positions def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] , lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): if is_complete(lowerCAmelCase_ ): return True for position in get_valid_pos(lowerCAmelCase_ , len(lowerCAmelCase_ ) ): snake_case_ ,snake_case_ : Dict = position if board[y][x] == 0: snake_case_ : List[str] = curr + 1 if open_knight_tour_helper(lowerCAmelCase_ , lowerCAmelCase_ , curr + 1 ): return True snake_case_ : Dict = 0 return False def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): snake_case_ : Optional[Any] = 1 if open_knight_tour_helper(lowerCAmelCase_ , (i, j) , 1 ): return board snake_case_ : Dict = 0 snake_case_ : str = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
666
0
"""simple docstring""" def __A ( a_ :int = 60_08_51_47_51_43) -> int: try: __a : Dict = int(a_) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''') if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''') __a : List[str] = 1 __a : List[str] = 2 while i * i <= n: while n % i == 0: __a : Dict = i n //= i i += 1 if n > 1: __a : Any = n return int(a_) if __name__ == "__main__": print(F'{solution() = }')
52
from ...configuration_utils import PretrainedConfig class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = "bert-generation" def __init__( self : Optional[int] , A__ : List[Any]=5_03_58 , A__ : Any=10_24 , A__ : Any=24 , A__ : List[Any]=16 , A__ : List[Any]=40_96 , A__ : int="gelu" , A__ : List[str]=0.1 , A__ : List[str]=0.1 , A__ : str=5_12 , A__ : int=0.02 , A__ : Any=1E-12 , A__ : Optional[Any]=0 , A__ : List[str]=2 , A__ : Optional[int]=1 , A__ : str="absolute" , A__ : Any=True , **A__ : Optional[Any] , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , **A__ ) snake_case_ : str = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Optional[Any] = hidden_act snake_case_ : Tuple = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : Optional[Any] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = position_embedding_type snake_case_ : Dict = use_cache
666
0
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _snake_case : Optional[int] = logging.getLogger(__name__) _snake_case : str = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _snake_case : Optional[int] = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _UpperCAmelCase : """simple docstring""" a_ = field( default=_UpperCamelCase , metadata={ """help""": ( """The model checkpoint for weights initialization. Leave None if you want to train a model from""" """ scratch.""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(_UpperCamelCase )} , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class _UpperCAmelCase : """simple docstring""" a_ = field( default=_UpperCamelCase , metadata={"""help""": """The input training data file (a text file)."""} ) a_ = field( default=_UpperCamelCase , metadata={ """help""": ( """The input training data files (multiple files in glob format). """ """Very often splitting large files to smaller files can prevent tokenizer going out of memory""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """An optional input train ref data file for whole word mask in Chinese."""} , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """An optional input eval ref data file for whole word mask in Chinese."""} , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Whether distinct lines of text in the dataset are to be handled as distinct sequences."""} , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Train with masked-language modeling loss instead of language modeling."""} ) a_ = field(default=_UpperCamelCase , metadata={"""help""": """Whether ot not to use whole word mask."""} ) a_ = field( default=0.15 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) a_ = field( default=1 / 6 , metadata={ """help""": ( """Ratio of length of a span of masked tokens to surrounding context length for permutation language""" """ modeling.""" ) } , ) a_ = field( default=5 , metadata={"""help""": """Maximum length of a span of masked tokens for permutation language modeling."""} ) a_ = field( default=-1 , metadata={ """help""": ( """Optional input sequence length after tokenization.""" """The training dataset will be truncated in block of this size for training.""" """Default to the model max input length for single sentence inputs (take into account special tokens).""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def a_ ( lowerCAmelCase_ : DataTrainingArguments, lowerCAmelCase_ : PreTrainedTokenizer, lowerCAmelCase_ : bool = False, lowerCAmelCase_ : Optional[str] = None, ): def _dataset(lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int]=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=lowerCAmelCase_, file_path=lowerCAmelCase_, block_size=args.block_size, ref_path=lowerCAmelCase_, ) return LineByLineTextDataset(tokenizer=lowerCAmelCase_, file_path=lowerCAmelCase_, block_size=args.block_size ) else: return TextDataset( tokenizer=lowerCAmelCase_, file_path=lowerCAmelCase_, block_size=args.block_size, overwrite_cache=args.overwrite_cache, cache_dir=lowerCAmelCase_, ) if evaluate: return _dataset(args.eval_data_file, args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(lowerCAmelCase_ ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file, args.train_ref_file ) def a_ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __lowerCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN, ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s', training_args.local_rank, training_args.device, training_args.n_gpu, bool(training_args.local_rank != -1 ), training_args.fpaa, ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s', lowerCAmelCase_ ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.config_name, cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path, cache_dir=model_args.cache_dir ) else: __lowerCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: __lowerCAmelCase = AutoTokenizer.from_pretrained(model_args.tokenizer_name, cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: __lowerCAmelCase = AutoTokenizer.from_pretrained(model_args.model_name_or_path, cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: __lowerCAmelCase = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path, from_tf=bool('.ckpt' in model_args.model_name_or_path ), config=lowerCAmelCase_, cache_dir=model_args.cache_dir, ) else: logger.info('Training new model from scratch' ) __lowerCAmelCase = AutoModelWithLMHead.from_config(lowerCAmelCase_ ) model.resize_token_embeddings(len(lowerCAmelCase_ ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: __lowerCAmelCase = tokenizer.max_len # Our input block size will be the max possible for the model else: __lowerCAmelCase = min(data_args.block_size, tokenizer.max_len ) # Get datasets __lowerCAmelCase = ( get_dataset(lowerCAmelCase_, tokenizer=lowerCAmelCase_, cache_dir=model_args.cache_dir ) if training_args.do_train else None ) __lowerCAmelCase = ( get_dataset(lowerCAmelCase_, tokenizer=lowerCAmelCase_, evaluate=lowerCAmelCase_, cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": __lowerCAmelCase = DataCollatorForPermutationLanguageModeling( tokenizer=lowerCAmelCase_, plm_probability=data_args.plm_probability, max_span_length=data_args.max_span_length, ) else: if data_args.mlm and data_args.whole_word_mask: __lowerCAmelCase = DataCollatorForWholeWordMask( tokenizer=lowerCAmelCase_, mlm_probability=data_args.mlm_probability ) else: __lowerCAmelCase = DataCollatorForLanguageModeling( tokenizer=lowerCAmelCase_, mlm=data_args.mlm, mlm_probability=data_args.mlm_probability ) # Initialize our Trainer __lowerCAmelCase = Trainer( model=lowerCAmelCase_, args=lowerCAmelCase_, data_collator=lowerCAmelCase_, train_dataset=lowerCAmelCase_, eval_dataset=lowerCAmelCase_, prediction_loss_only=lowerCAmelCase_, ) # Training if training_args.do_train: __lowerCAmelCase = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=lowerCAmelCase_ ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation __lowerCAmelCase = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) __lowerCAmelCase = trainer.evaluate() __lowerCAmelCase = math.exp(eval_output['eval_loss'] ) __lowerCAmelCase = {'perplexity': perplexity} __lowerCAmelCase = os.path.join(training_args.output_dir, 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(lowerCAmelCase_, 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s', lowerCAmelCase_, str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(lowerCAmelCase_ ) return results def a_ ( lowerCAmelCase_ : Optional[Any] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
53
import math def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [] snake_case_ : List[str] = 2 snake_case_ : Optional[int] = int(math.sqrt(lowerCAmelCase_ ) ) # Size of every segment snake_case_ : str = [True] * (end + 1) snake_case_ : Any = [] while start <= end: if temp[start] is True: in_prime.append(lowerCAmelCase_ ) for i in range(start * start , end + 1 , lowerCAmelCase_ ): snake_case_ : Union[str, Any] = False start += 1 prime += in_prime snake_case_ : Dict = end + 1 snake_case_ : Dict = min(2 * end , lowerCAmelCase_ ) while low <= n: snake_case_ : Any = [True] * (high - low + 1) for each in in_prime: snake_case_ : Optional[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(lowerCAmelCase_ , high + 1 , lowerCAmelCase_ ): snake_case_ : List[Any] = False for j in range(len(lowerCAmelCase_ ) ): if temp[j] is True: prime.append(j + low ) snake_case_ : int = high + 1 snake_case_ : Union[str, Any] = min(high + end , lowerCAmelCase_ ) return prime print(sieve(1_0**6))
666
0
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A : def __init__( self: List[str] , _lowerCAmelCase: Tuple , _lowerCAmelCase: Dict=13 , _lowerCAmelCase: Optional[Any]=32 , _lowerCAmelCase: Optional[int]=3 , _lowerCAmelCase: List[str]=4 , _lowerCAmelCase: Any=[10, 20, 30, 40] , _lowerCAmelCase: Tuple=[2, 2, 3, 2] , _lowerCAmelCase: Tuple=True , _lowerCAmelCase: Any=True , _lowerCAmelCase: Union[str, Any]=37 , _lowerCAmelCase: Optional[Any]="gelu" , _lowerCAmelCase: Optional[Any]=10 , _lowerCAmelCase: Optional[Any]=0.02 , _lowerCAmelCase: List[str]=["stage2", "stage3", "stage4"] , _lowerCAmelCase: List[Any]=[2, 3, 4] , _lowerCAmelCase: Any=None , ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ =parent UpperCAmelCase_ =batch_size UpperCAmelCase_ =image_size UpperCAmelCase_ =num_channels UpperCAmelCase_ =num_stages UpperCAmelCase_ =hidden_sizes UpperCAmelCase_ =depths UpperCAmelCase_ =is_training UpperCAmelCase_ =use_labels UpperCAmelCase_ =intermediate_size UpperCAmelCase_ =hidden_act UpperCAmelCase_ =num_labels UpperCAmelCase_ =initializer_range UpperCAmelCase_ =out_features UpperCAmelCase_ =out_indices UpperCAmelCase_ =scope def lowerCAmelCase__ ( self: Optional[int] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ =None if self.use_labels: UpperCAmelCase_ =ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ =self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self: int ) -> List[str]: '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def lowerCAmelCase__ ( self: List[Any] , _lowerCAmelCase: Dict , _lowerCAmelCase: List[Any] , _lowerCAmelCase: List[Any] ) -> str: '''simple docstring''' UpperCAmelCase_ =ConvNextModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase_ =model(_lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCAmelCase__ ( self: int , _lowerCAmelCase: List[str] , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Any ) -> int: '''simple docstring''' UpperCAmelCase_ =ConvNextForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase_ =model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self: int , _lowerCAmelCase: List[str] , _lowerCAmelCase: Dict , _lowerCAmelCase: List[Any] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ =ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase_ =model(_lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCAmelCase_ =None UpperCAmelCase_ =ConvNextBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase_ =model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowerCAmelCase__ ( self: Optional[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ =self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ =config_and_inputs UpperCAmelCase_ ={"pixel_values": pixel_values} return config, inputs_dict @require_torch class A ( __lowercase , __lowercase , unittest.TestCase ): _snake_case =( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) _snake_case =( {'''feature-extraction''': ConvNextModel, '''image-classification''': ConvNextForImageClassification} if is_torch_available() else {} ) _snake_case =True _snake_case =False _snake_case =False _snake_case =False _snake_case =False def lowerCAmelCase__ ( self: Dict ) -> Dict: '''simple docstring''' UpperCAmelCase_ =ConvNextModelTester(self ) UpperCAmelCase_ =ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def lowerCAmelCase__ ( self: str ) -> str: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self: int ) -> str: '''simple docstring''' return @unittest.skip(reason="ConvNext does not use inputs_embeds" ) def lowerCAmelCase__ ( self: Union[str, Any] ) -> List[str]: '''simple docstring''' pass @unittest.skip(reason="ConvNext does not support input and output embeddings" ) def lowerCAmelCase__ ( self: Dict ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="ConvNext does not use feedforward chunking" ) def lowerCAmelCase__ ( self: List[Any] ) -> Any: '''simple docstring''' pass def lowerCAmelCase__ ( self: Union[str, Any] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ =model_class(_lowerCAmelCase ) UpperCAmelCase_ =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ =[*signature.parameters.keys()] UpperCAmelCase_ =["pixel_values"] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def lowerCAmelCase__ ( self: Any ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: List[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: List[Any] ) -> Any: '''simple docstring''' def check_hidden_states_output(_lowerCAmelCase: List[Any] , _lowerCAmelCase: Optional[Any] , _lowerCAmelCase: int ): UpperCAmelCase_ =model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ =model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase_ =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase_ =self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCAmelCase_ , UpperCAmelCase_ =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ =True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ =True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase__ ( self: str ) -> Tuple: '''simple docstring''' UpperCAmelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def lowerCAmelCase__ ( self: str ) -> List[str]: '''simple docstring''' for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ =ConvNextModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def a__ ( ): '''simple docstring''' UpperCAmelCase_ =Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A ( unittest.TestCase ): @cached_property def lowerCAmelCase__ ( self: Tuple ) -> List[Any]: '''simple docstring''' return AutoImageProcessor.from_pretrained("facebook/convnext-tiny-224" ) if is_vision_available() else None @slow def lowerCAmelCase__ ( self: List[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ =ConvNextForImageClassification.from_pretrained("facebook/convnext-tiny-224" ).to(_lowerCAmelCase ) UpperCAmelCase_ =self.default_image_processor UpperCAmelCase_ =prepare_img() UpperCAmelCase_ =image_processor(images=_lowerCAmelCase , return_tensors="pt" ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ =model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase_ =torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase_ =torch.tensor([-0.02_60, -0.47_39, 0.19_11] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 ) ) @require_torch class A ( unittest.TestCase , __lowercase ): _snake_case =(ConvNextBackbone,) if is_torch_available() else () _snake_case =ConvNextConfig _snake_case =False def lowerCAmelCase__ ( self: List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase_ =ConvNextModelTester(self )
54
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self : List[str] , A__ : List[Any] , A__ : int=7 , A__ : Union[str, Any]=3 , A__ : List[str]=30 , A__ : Optional[int]=4_00 , A__ : Optional[Any]=True , A__ : Optional[int]=None , A__ : Optional[Any]=True , A__ : Any=[0.5, 0.5, 0.5] , A__ : int=[0.5, 0.5, 0.5] , A__ : Any=True , A__ : int=1 / 2_55 , A__ : List[str]=True , ) -> Dict: '''simple docstring''' snake_case_ : int = size if size is not None else {"shortest_edge": 18, "longest_edge": 13_33} snake_case_ : Any = parent snake_case_ : Optional[int] = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Union[str, Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Tuple = do_resize snake_case_ : Dict = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : List[Any] = image_std snake_case_ : Tuple = do_rescale snake_case_ : Any = rescale_factor snake_case_ : Optional[int] = do_pad def UpperCAmelCase__ ( self : int ) -> List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase__ ( self : Optional[int] , A__ : Optional[int] , A__ : Any=False ) -> Optional[Any]: '''simple docstring''' if not batched: snake_case_ : Any = image_inputs[0] if isinstance(A__ , Image.Image ): snake_case_ ,snake_case_ : Dict = image.size else: snake_case_ ,snake_case_ : int = image.shape[1], image.shape[2] if w < h: snake_case_ : Dict = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[int] = self.size["shortest_edge"] elif w > h: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : str = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : List[Any] = self.size["shortest_edge"] else: snake_case_ : str = [] for image in image_inputs: snake_case_ ,snake_case_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : List[Any] = max(A__ , key=lambda A__ : item[0] )[0] snake_case_ : int = max(A__ , key=lambda A__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( _UpperCamelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[int] = ConditionalDetrImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' snake_case_ : List[str] = ConditionalDetrImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Any ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A__ , "image_mean" ) ) self.assertTrue(hasattr(A__ , "image_std" ) ) self.assertTrue(hasattr(A__ , "do_normalize" ) ) self.assertTrue(hasattr(A__ , "do_resize" ) ) self.assertTrue(hasattr(A__ , "size" ) ) def UpperCAmelCase__ ( self : List[str] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 13_33} ) self.assertEqual(image_processor.do_pad , A__ ) snake_case_ : Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A__ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , A__ ) def UpperCAmelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' pass def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ ) for image in image_inputs: self.assertIsInstance(A__ , Image.Image ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) snake_case_ : int = image_processing(A__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : int ) -> Any: '''simple docstring''' snake_case_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , numpify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , np.ndarray ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[str] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Optional[int] = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Dict = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : Tuple ) -> str: '''simple docstring''' snake_case_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , torchify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , torch.Tensor ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Any = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : int = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def UpperCAmelCase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : Optional[Any] = json.loads(f.read() ) snake_case_ : int = {"image_id": 3_97_69, "annotations": target} # encode them snake_case_ : Optional[int] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) snake_case_ : Any = image_processing(images=A__ , annotations=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : List[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : List[str] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Tuple = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify orig_size snake_case_ : Any = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) ) @slow def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' snake_case_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : Any = json.loads(f.read() ) snake_case_ : Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_97_69, "segments_info": target} snake_case_ : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor(format="coco_panoptic" ) snake_case_ : str = image_processing(images=A__ , annotations=A__ , masks_path=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : int = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Optional[int] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Optional[int] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify masks snake_case_ : Union[str, Any] = 82_28_73 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , A__ ) # verify orig_size snake_case_ : Dict = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : str = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) )
666
0
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE :int = get_tests_dir('fixtures/test_sentencepiece_no_bos.model') @require_sentencepiece @require_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = PegasusTokenizer snake_case_ = PegasusTokenizerFast snake_case_ = True snake_case_ = True def UpperCamelCase_ ( self : Dict ): super().setUp() # We have a SentencePiece fixture for testing __A = PegasusTokenizer(A ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase_ ( self : Union[str, Any] ): return PegasusTokenizer.from_pretrained("google/pegasus-large" ) def UpperCamelCase_ ( self : int ,**A : Dict ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def UpperCamelCase_ ( self : List[str] ,A : str ): return ("This is a test", "This is a test") def UpperCamelCase_ ( self : Any ): __A = "</s>" __A = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def UpperCamelCase_ ( self : List[str] ): __A = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"<pad>" ) self.assertEqual(vocab_keys[1] ,"</s>" ) self.assertEqual(vocab_keys[-1] ,"v" ) self.assertEqual(len(A ) ,11_03 ) def UpperCamelCase_ ( self : Union[str, Any] ): self.assertEqual(self.get_tokenizer().vocab_size ,11_03 ) def UpperCamelCase_ ( self : int ): __A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A = ( "Let's see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important" " </s> <pad> <pad> <pad>" ) __A = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] __A = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : Optional[int] ): __A = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word __A = "<mask_1> To ensure a <mask_2> flow of bank resolutions." __A = [2, 4_13, 6_15, 1_14, 3, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] __A = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : Tuple ): __A = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_61_03 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_03 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_05 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 10_24 __A = "To ensure a smooth flow of bank resolutions." __A = [4_13, 6_15, 1_14, 22_91, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] __A = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def UpperCamelCase_ ( self : Union[str, Any] ): __A = ["This is going to be way too long." * 1_50, "short example"] __A = ["not super long but more than 5 tokens", "tiny"] __A = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="pt" ) __A = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="pt" ) assert batch.input_ids.shape == (2, 10_24) assert batch.attention_mask.shape == (2, 10_24) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. @slow def UpperCamelCase_ ( self : List[str] ): # fmt: off __A = {"input_ids": [[3_89_79, 1_43, 1_84_85, 6_06, 1_30, 2_66_69, 8_76_86, 1_21, 5_41_89, 11_29, 1_11, 2_66_69, 8_76_86, 1_21, 91_14, 1_47_87, 1_21, 1_32_49, 1_58, 5_92, 9_56, 1_21, 1_46_21, 3_15_76, 1_43, 6_26_13, 1_08, 96_88, 9_30, 4_34_30, 1_15_62, 6_26_13, 3_04, 1_08, 1_14_43, 8_97, 1_08, 93_14, 1_74_15, 6_33_99, 1_08, 1_14_43, 76_14, 1_83_16, 1_18, 42_84, 71_48, 1_24_30, 1_43, 14_00, 2_57_03, 1_58, 1_11, 42_84, 71_48, 1_17_72, 1_43, 2_12_97, 10_64, 1_58, 1_22, 2_04, 35_06, 17_54, 11_33, 1_47_87, 15_81, 1_15, 3_32_24, 44_82, 1_11, 13_55, 1_10, 2_91_73, 3_17, 5_08_33, 1_08, 2_01_47, 9_46_65, 1_11, 7_71_98, 1_07, 1], [1_10, 6_26_13, 1_17, 6_38, 1_12, 11_33, 1_21, 2_00_98, 13_55, 7_90_50, 1_38_72, 1_35, 15_96, 5_35_41, 13_52, 1_41, 1_30_39, 55_42, 1_24, 3_02, 5_18, 1_11, 2_68, 29_56, 1_15, 1_49, 44_27, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_39, 12_35, 27_99, 1_82_89, 1_77_80, 2_04, 1_09, 94_74, 12_96, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="google/bigbird-pegasus-large-arxiv" ,revision="ba85d0851d708441f91440d509690f1ab6353415" ,) @require_sentencepiece @require_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = PegasusTokenizer snake_case_ = PegasusTokenizerFast snake_case_ = True snake_case_ = True def UpperCamelCase_ ( self : Tuple ): super().setUp() # We have a SentencePiece fixture for testing __A = PegasusTokenizer(A ,offset=0 ,mask_token_sent=A ,mask_token="[MASK]" ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase_ ( self : Union[str, Any] ): return PegasusTokenizer.from_pretrained("google/bigbird-pegasus-large-arxiv" ) def UpperCamelCase_ ( self : str ,**A : Union[str, Any] ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def UpperCamelCase_ ( self : Optional[int] ,A : Dict ): return ("This is a test", "This is a test") def UpperCamelCase_ ( self : int ): __A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A = ( "Let's see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>" " <pad> <pad> <pad>" ) __A = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] __A = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) @require_torch def UpperCamelCase_ ( self : Optional[Any] ): __A = ["This is going to be way too long." * 10_00, "short example"] __A = ["not super long but more than 5 tokens", "tiny"] __A = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="pt" ) __A = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="pt" ) assert batch.input_ids.shape == (2, 40_96) assert batch.attention_mask.shape == (2, 40_96) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. def UpperCamelCase_ ( self : Dict ): __A = ( "This is an example string that is used to test the original TF implementation against the HF" " implementation" ) __A = self._large_tokenizer(A ).input_ids self.assertListEqual( A ,[1_82, 1_17, 1_42, 5_87, 42_11, 1_20, 1_17, 2_63, 1_12, 8_04, 1_09, 8_56, 2_50_16, 31_37, 4_64, 1_09, 2_69_55, 31_37, 1] ,)
55
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "Model type selected in the list: " + ", ".join(_UpperCamelCase )} ) _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) _SCREAMING_SNAKE_CASE : int = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) _SCREAMING_SNAKE_CASE : float = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) _SCREAMING_SNAKE_CASE : int = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = "train" _SCREAMING_SNAKE_CASE : Any = "dev" class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : SquadDataTrainingArguments _SCREAMING_SNAKE_CASE : List[SquadFeatures] _SCREAMING_SNAKE_CASE : Split _SCREAMING_SNAKE_CASE : bool def __init__( self : str , A__ : SquadDataTrainingArguments , A__ : PreTrainedTokenizer , A__ : Optional[int] = None , A__ : Union[str, Split] = Split.train , A__ : Optional[bool] = False , A__ : Optional[str] = None , A__ : Optional[str] = "pt" , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = args snake_case_ : int = is_language_sensitive snake_case_ : int = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(A__ , A__ ): try: snake_case_ : List[str] = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) snake_case_ : Tuple = mode # Load data features from cache or dataset file snake_case_ : Dict = "v2" if args.version_2_with_negative else "v1" snake_case_ : List[Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case_ : List[Any] = cached_features_file + ".lock" with FileLock(A__ ): if os.path.exists(A__ ) and not args.overwrite_cache: snake_case_ : int = time.time() snake_case_ : List[Any] = torch.load(A__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. snake_case_ : Tuple = self.old_features["features"] snake_case_ : List[str] = self.old_features.get("dataset" , A__ ) snake_case_ : Tuple = self.old_features.get("examples" , A__ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" " future run" ) else: if mode == Split.dev: snake_case_ : Tuple = self.processor.get_dev_examples(args.data_dir ) else: snake_case_ : Tuple = self.processor.get_train_examples(args.data_dir ) snake_case_ ,snake_case_ : Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=A__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=A__ , ) snake_case_ : Any = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , A__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self : str ) -> Dict: '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[int] , A__ : Optional[int] ) -> Dict[str, torch.Tensor]: '''simple docstring''' snake_case_ : Any = self.features[i] snake_case_ : Optional[int] = torch.tensor(feature.input_ids , dtype=torch.long ) snake_case_ : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.cls_index , dtype=torch.long ) snake_case_ : str = torch.tensor(feature.p_mask , dtype=torch.float ) snake_case_ : str = torch.tensor(feature.is_impossible , dtype=torch.float ) snake_case_ : Optional[int] = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: snake_case_ : Any = torch.tensor(feature.start_position , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
666
0
'''simple docstring''' import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _a : str = ( "4S 3H 2C 7S 5H", "9D 8H 2C 6S 7H", "2D 6D 9D TH 7D", "TC 8C 2S JH 6C", "JH 8S TH AH QH", "TS KS 5S 9S AC", "KD 6S 9D TH AD", "KS 8D 4D 9S 4S", # pair "8C 4S KH JS 4D", # pair "QH 8H KD JH 8S", # pair "KC 4H KS 2H 8D", # pair "KD 4S KC 3H 8S", # pair "AH 8S AS KC JH", # pair "3H 4C 4H 3S 2H", # 2 pairs "5S 5D 2C KH KH", # 2 pairs "3C KH 5D 5S KH", # 2 pairs "AS 3C KH AD KH", # 2 pairs "7C 7S 3S 7H 5S", # 3 of a kind "7C 7S KH 2H 7H", # 3 of a kind "AC KH QH AH AS", # 3 of a kind "2H 4D 3C AS 5S", # straight (low ace) "3C 5C 4C 2C 6H", # straight "6S 8S 7S 5H 9H", # straight "JS QS 9H TS KH", # straight "QC KH TS JS AH", # straight (high ace) "8C 9C 5C 3C TC", # flush "3S 8S 9S 5S KS", # flush "4C 5C 9C 8C KC", # flush "JH 8H AH KH QH", # flush "3D 2H 3H 2C 2D", # full house "2H 2C 3S 3H 3D", # full house "KH KC 3S 3H 3D", # full house "JC 6H JS JD JH", # 4 of a kind "JC 7H JS JD JH", # 4 of a kind "JC KH JS JD JH", # 4 of a kind "2S AS 4S 5S 3S", # straight flush (low ace) "2D 6D 3D 4D 5D", # straight flush "5C 6C 3C 7C 4C", # straight flush "JH 9H TH KH QH", # straight flush "JH AH TH KH QH", # royal flush (high ace straight flush) ) _a : str = ( ("2H 3H 4H 5H 6H", "KS AS TS QS JS", "Loss"), ("2H 3H 4H 5H 6H", "AS AD AC AH JD", "Win"), ("AS AH 2H AD AC", "JS JD JC JH 3D", "Win"), ("2S AH 2H AS AC", "JS JD JC JH AD", "Loss"), ("2S AH 2H AS AC", "2H 3H 5H 6H 7H", "Win"), ("AS 3S 4S 8S 2S", "2H 3H 5H 6H 7H", "Win"), ("2H 3H 5H 6H 7H", "2S 3H 4H 5S 6C", "Win"), ("2S 3H 4H 5S 6C", "3D 4C 5H 6H 2S", "Tie"), ("2S 3H 4H 5S 6C", "AH AC 5H 6H AS", "Win"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H AS", "Loss"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H 7S", "Win"), ("6S AD 7H 4S AS", "AH AC 5H 6H 7S", "Loss"), ("2S AH 4H 5S KC", "AH AC 5H 6H 7S", "Loss"), ("2S 3H 6H 7S 9C", "7H 3C TH 6H 9S", "Loss"), ("4S 5H 6H TS AC", "3S 5H 6H TS AC", "Win"), ("2S AH 4H 5S 6C", "AD 4C 5H 6H 2C", "Tie"), ("AS AH 3H AD AC", "AS AH 2H AD AC", "Win"), ("AH AC 5H 5C QS", "AH AC 5H 5C KS", "Loss"), ("AH AC 5H 5C QS", "KH KC 5H 5C QS", "Win"), ("7C 7S KH 2H 7H", "3C 3S AH 2H 3H", "Win"), ("3C 3S AH 2H 3H", "7C 7S KH 2H 7H", "Loss"), ("6H 5H 4H 3H 2H", "5H 4H 3H 2H AH", "Win"), ("5H 4H 3H 2H AH", "5H 4H 3H 2H AH", "Tie"), ("5H 4H 3H 2H AH", "6H 5H 4H 3H 2H", "Loss"), ("AH AD KS KC AC", "AH KD KH AC KC", "Win"), ("2H 4D 3C AS 5S", "2H 4D 3C 6S 5S", "Loss"), ("2H 3S 3C 3H 2S", "3S 3C 2S 2H 2D", "Win"), ("4D 6D 5D 2D JH", "3S 8S 3H TC KH", "Loss"), ("4S 6C 8S 3S 7S", "AD KS 2D 7D 7C", "Loss"), ("6S 4C 7H 8C 3H", "5H JC AH 9D 9C", "Loss"), ("9D 9H JH TC QH", "3C 2S JS 5C 7H", "Win"), ("2H TC 8S AD 9S", "4H TS 7H 2C 5C", "Win"), ("9D 3S 2C 7S 7C", "JC TD 3C TC 9H", "Loss"), ) _a : int = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", True), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", False), ("AS 3S 4S 8S 2S", True), ) _a : int = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", False), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", True), ) _a : Optional[int] = ( ("2H 4D 3C AS 5S", True, [5, 4, 3, 2, 14]), ("2H 5D 3C AS 5S", False, [14, 5, 5, 3, 2]), ("JH QD KC AS TS", False, [14, 13, 12, 11, 10]), ("9D 3S 2C 7S 7C", False, [9, 7, 7, 3, 2]), ) _a : Union[str, Any] = ( ("JH AH TH KH QH", 0), ("JH 9H TH KH QH", 0), ("JC KH JS JD JH", 7), ("KH KC 3S 3H 3D", 6), ("8C 9C 5C 3C TC", 0), ("JS QS 9H TS KH", 0), ("7C 7S KH 2H 7H", 3), ("3C KH 5D 5S KH", 2), ("QH 8H KD JH 8S", 1), ("2D 6D 9D TH 7D", 0), ) _a : Any = ( ("JH AH TH KH QH", 23), ("JH 9H TH KH QH", 22), ("JC KH JS JD JH", 21), ("KH KC 3S 3H 3D", 20), ("8C 9C 5C 3C TC", 19), ("JS QS 9H TS KH", 18), ("7C 7S KH 2H 7H", 17), ("3C KH 5D 5S KH", 16), ("QH 8H KD JH 8S", 15), ("2D 6D 9D TH 7D", 14), ) def _a () -> str: """simple docstring""" __snake_case , __snake_case = randrange(len(lowercase__ ) ), randrange(len(lowercase__ ) ) __snake_case = ['Loss', 'Tie', 'Win'][(play >= oppo) + (play > oppo)] __snake_case , __snake_case = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def _a (lowercase__ : int = 1_0_0 ) -> Optional[int]: """simple docstring""" return (generate_random_hand() for _ in range(lowercase__ )) @pytest.mark.parametrize('hand, expected' , lowercase__ ) def _a (lowercase__ : Optional[int] , lowercase__ : Any ) -> Dict: """simple docstring""" assert PokerHand(lowercase__ )._is_flush() == expected @pytest.mark.parametrize('hand, expected' , lowercase__ ) def _a (lowercase__ : int , lowercase__ : List[Any] ) -> List[Any]: """simple docstring""" assert PokerHand(lowercase__ )._is_straight() == expected @pytest.mark.parametrize('hand, expected, card_values' , lowercase__ ) def _a (lowercase__ : int , lowercase__ : Tuple , lowercase__ : List[str] ) -> str: """simple docstring""" __snake_case = PokerHand(lowercase__ ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('hand, expected' , lowercase__ ) def _a (lowercase__ : str , lowercase__ : int ) -> Any: """simple docstring""" assert PokerHand(lowercase__ )._is_same_kind() == expected @pytest.mark.parametrize('hand, expected' , lowercase__ ) def _a (lowercase__ : List[str] , lowercase__ : str ) -> Union[str, Any]: """simple docstring""" assert PokerHand(lowercase__ )._hand_type == expected @pytest.mark.parametrize('hand, other, expected' , lowercase__ ) def _a (lowercase__ : Union[str, Any] , lowercase__ : Tuple , lowercase__ : Union[str, Any] ) -> str: """simple docstring""" assert PokerHand(lowercase__ ).compare_with(PokerHand(lowercase__ ) ) == expected @pytest.mark.parametrize('hand, other, expected' , generate_random_hands() ) def _a (lowercase__ : int , lowercase__ : int , lowercase__ : List[str] ) -> str: """simple docstring""" assert PokerHand(lowercase__ ).compare_with(PokerHand(lowercase__ ) ) == expected def _a () -> Any: """simple docstring""" __snake_case = [PokerHand(lowercase__ ) for hand in SORTED_HANDS] __snake_case = poker_hands.copy() shuffle(lowercase__ ) __snake_case = chain(sorted(lowercase__ ) ) for index, hand in enumerate(lowercase__ ): assert hand == poker_hands[index] def _a () -> str: """simple docstring""" # Test that five high straights are compared correctly. __snake_case = [PokerHand('2D AC 3H 4H 5S' ), PokerHand('2S 3H 4H 5S 6C' )] pokerhands.sort(reverse=lowercase__ ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def _a () -> Dict: """simple docstring""" # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. __snake_case = PokerHand('2C 4S AS 3D 5C' ) __snake_case = True __snake_case = [5, 4, 3, 2, 1_4] for _ in range(1_0 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def _a () -> Any: """simple docstring""" # Problem number 54 from Project Euler # Testing from poker_hands.txt file __snake_case = 0 __snake_case = os.path.abspath(os.path.dirname(lowercase__ ) ) __snake_case = os.path.join(lowercase__ , 'poker_hands.txt' ) with open(lowercase__ ) as file_hand: for line in file_hand: __snake_case = line[:1_4].strip() __snake_case = line[1_5:].strip() __snake_case , __snake_case = PokerHand(lowercase__ ), PokerHand(lowercase__ ) __snake_case = player.compare_with(lowercase__ ) if output == "Win": answer += 1 assert answer == 3_7_6
56
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "git_vision_model" def __init__( self : int , A__ : Union[str, Any]=7_68 , A__ : List[Any]=30_72 , A__ : Tuple=12 , A__ : Optional[Any]=12 , A__ : Optional[int]=3 , A__ : List[str]=2_24 , A__ : Dict=16 , A__ : int="quick_gelu" , A__ : Any=1E-5 , A__ : Tuple=0.0 , A__ : Optional[int]=0.02 , **A__ : List[str] , ) -> Optional[int]: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[Any] = hidden_size snake_case_ : str = intermediate_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = num_channels snake_case_ : Union[str, Any] = patch_size snake_case_ : List[str] = image_size snake_case_ : List[Any] = initializer_range snake_case_ : Any = attention_dropout snake_case_ : Any = layer_norm_eps snake_case_ : int = hidden_act @classmethod def UpperCAmelCase__ ( cls : List[Any] , A__ : Union[str, os.PathLike] , **A__ : Optional[int] ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(A__ ) snake_case_ ,snake_case_ : Tuple = cls.get_config_dict(A__ , **A__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": snake_case_ : Any = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(A__ , **A__ ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = "git" def __init__( self : Any , A__ : List[str]=None , A__ : List[str]=3_05_22 , A__ : Tuple=7_68 , A__ : Tuple=6 , A__ : str=12 , A__ : Any=30_72 , A__ : List[str]="gelu" , A__ : int=0.1 , A__ : Dict=0.1 , A__ : Any=10_24 , A__ : Optional[Any]=0.02 , A__ : Optional[Any]=1E-12 , A__ : Dict=0 , A__ : Any="absolute" , A__ : Tuple=True , A__ : Any=False , A__ : Tuple=1_01 , A__ : Tuple=1_02 , A__ : List[Any]=None , **A__ : List[str] , ) -> int: '''simple docstring''' super().__init__(bos_token_id=A__ , eos_token_id=A__ , pad_token_id=A__ , **A__ ) if vision_config is None: snake_case_ : int = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) snake_case_ : str = GitVisionConfig(**A__ ) snake_case_ : int = vocab_size snake_case_ : List[Any] = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : List[Any] = num_attention_heads snake_case_ : Any = hidden_act snake_case_ : Dict = intermediate_size snake_case_ : Any = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = initializer_range snake_case_ : List[str] = layer_norm_eps snake_case_ : Any = position_embedding_type snake_case_ : Union[str, Any] = use_cache snake_case_ : str = tie_word_embeddings snake_case_ : List[Any] = num_image_with_embedding snake_case_ : Dict = bos_token_id snake_case_ : int = eos_token_id def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' snake_case_ : Tuple = copy.deepcopy(self.__dict__ ) snake_case_ : Optional[int] = self.vision_config.to_dict() snake_case_ : Tuple = self.__class__.model_type return output
666
0
from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class _lowerCAmelCase( UpperCAmelCase_ ): """simple docstring""" a : "DiagonalGaussianDistribution" class _lowerCAmelCase( UpperCAmelCase_ , UpperCAmelCase_ ): """simple docstring""" a : str =True @register_to_config def __init__( self , _lowerCamelCase = 3 , _lowerCamelCase = 3 , _lowerCamelCase = ("DownEncoderBlock2D",) , _lowerCamelCase = ("UpDecoderBlock2D",) , _lowerCamelCase = (6_4,) , _lowerCamelCase = 1 , _lowerCamelCase = "silu" , _lowerCamelCase = 4 , _lowerCamelCase = 3_2 , _lowerCamelCase = 3_2 , _lowerCamelCase = 0.1_8_2_1_5 , ): super().__init__() # pass init params to Encoder UpperCamelCase_: List[Any] = Encoder( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , down_block_types=_lowerCamelCase , block_out_channels=_lowerCamelCase , layers_per_block=_lowerCamelCase , act_fn=_lowerCamelCase , norm_num_groups=_lowerCamelCase , double_z=_lowerCamelCase , ) # pass init params to Decoder UpperCamelCase_: Optional[Any] = Decoder( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , up_block_types=_lowerCamelCase , block_out_channels=_lowerCamelCase , layers_per_block=_lowerCamelCase , norm_num_groups=_lowerCamelCase , act_fn=_lowerCamelCase , ) UpperCamelCase_: Optional[Any] = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) UpperCamelCase_: Any = nn.Convad(_lowerCamelCase , _lowerCamelCase , 1 ) UpperCamelCase_: List[str] = False UpperCamelCase_: Tuple = False # only relevant if vae tiling is enabled UpperCamelCase_: Union[str, Any] = self.config.sample_size UpperCamelCase_: Dict = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) UpperCamelCase_: str = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) UpperCamelCase_: str = 0.2_5 def _a ( self , _lowerCamelCase , _lowerCamelCase=False ): if isinstance(_lowerCamelCase , (Encoder, Decoder) ): UpperCamelCase_: str = value def _a ( self , _lowerCamelCase = True ): UpperCamelCase_: Any = use_tiling def _a ( self ): self.enable_tiling(_lowerCamelCase ) def _a ( self ): UpperCamelCase_: List[Any] = True def _a ( self ): UpperCamelCase_: Optional[Any] = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def _a ( self ): UpperCamelCase_: Dict = {} def fn_recursive_add_processors(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): if hasattr(_lowerCamelCase , 'set_processor' ): UpperCamelCase_: List[Any] = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , _lowerCamelCase , _lowerCamelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) return processors def _a ( self , _lowerCamelCase ): UpperCamelCase_: Tuple = len(self.attn_processors.keys() ) if isinstance(_lowerCamelCase , _lowerCamelCase ) and len(_lowerCamelCase ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(_lowerCamelCase )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): if hasattr(_lowerCamelCase , 'set_processor' ): if not isinstance(_lowerCamelCase , _lowerCamelCase ): module.set_processor(_lowerCamelCase ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , _lowerCamelCase , _lowerCamelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a ( self ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def _a ( self , _lowerCamelCase , _lowerCamelCase = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(_lowerCamelCase , return_dict=_lowerCamelCase ) if self.use_slicing and x.shape[0] > 1: UpperCamelCase_: List[Any] = [self.encoder(_lowerCamelCase ) for x_slice in x.split(1 )] UpperCamelCase_: str = torch.cat(_lowerCamelCase ) else: UpperCamelCase_: Optional[Any] = self.encoder(_lowerCamelCase ) UpperCamelCase_: Optional[int] = self.quant_conv(_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = DiagonalGaussianDistribution(_lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_lowerCamelCase ) def _a ( self , _lowerCamelCase , _lowerCamelCase = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(_lowerCamelCase , return_dict=_lowerCamelCase ) UpperCamelCase_: Dict = self.post_quant_conv(_lowerCamelCase ) UpperCamelCase_: List[Any] = self.decoder(_lowerCamelCase ) if not return_dict: return (dec,) return DecoderOutput(sample=_lowerCamelCase ) @apply_forward_hook def _a ( self , _lowerCamelCase , _lowerCamelCase = True ): if self.use_slicing and z.shape[0] > 1: UpperCamelCase_: Optional[int] = [self._decode(_lowerCamelCase ).sample for z_slice in z.split(1 )] UpperCamelCase_: Optional[Any] = torch.cat(_lowerCamelCase ) else: UpperCamelCase_: Optional[int] = self._decode(_lowerCamelCase ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=_lowerCamelCase ) def _a ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): UpperCamelCase_: Optional[Any] = min(a.shape[2] , b.shape[2] , _lowerCamelCase ) for y in range(_lowerCamelCase ): UpperCamelCase_: Dict = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def _a ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): UpperCamelCase_: List[Any] = min(a.shape[3] , b.shape[3] , _lowerCamelCase ) for x in range(_lowerCamelCase ): UpperCamelCase_: List[Any] = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def _a ( self , _lowerCamelCase , _lowerCamelCase = True ): UpperCamelCase_: List[str] = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) UpperCamelCase_: Tuple = int(self.tile_latent_min_size * self.tile_overlap_factor ) UpperCamelCase_: int = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. UpperCamelCase_: Any = [] for i in range(0 , x.shape[2] , _lowerCamelCase ): UpperCamelCase_: Union[str, Any] = [] for j in range(0 , x.shape[3] , _lowerCamelCase ): UpperCamelCase_: int = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] UpperCamelCase_: int = self.encoder(_lowerCamelCase ) UpperCamelCase_: Union[str, Any] = self.quant_conv(_lowerCamelCase ) row.append(_lowerCamelCase ) rows.append(_lowerCamelCase ) UpperCamelCase_: Optional[int] = [] for i, row in enumerate(_lowerCamelCase ): UpperCamelCase_: List[Any] = [] for j, tile in enumerate(_lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: UpperCamelCase_: int = self.blend_v(rows[i - 1][j] , _lowerCamelCase , _lowerCamelCase ) if j > 0: UpperCamelCase_: List[Any] = self.blend_h(row[j - 1] , _lowerCamelCase , _lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(_lowerCamelCase , dim=3 ) ) UpperCamelCase_: Any = torch.cat(_lowerCamelCase , dim=2 ) UpperCamelCase_: Any = DiagonalGaussianDistribution(_lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_lowerCamelCase ) def _a ( self , _lowerCamelCase , _lowerCamelCase = True ): UpperCamelCase_: int = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) UpperCamelCase_: str = int(self.tile_sample_min_size * self.tile_overlap_factor ) UpperCamelCase_: Any = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. UpperCamelCase_: Optional[int] = [] for i in range(0 , z.shape[2] , _lowerCamelCase ): UpperCamelCase_: List[Any] = [] for j in range(0 , z.shape[3] , _lowerCamelCase ): UpperCamelCase_: int = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] UpperCamelCase_: List[str] = self.post_quant_conv(_lowerCamelCase ) UpperCamelCase_: List[str] = self.decoder(_lowerCamelCase ) row.append(_lowerCamelCase ) rows.append(_lowerCamelCase ) UpperCamelCase_: Optional[int] = [] for i, row in enumerate(_lowerCamelCase ): UpperCamelCase_: int = [] for j, tile in enumerate(_lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: UpperCamelCase_: Any = self.blend_v(rows[i - 1][j] , _lowerCamelCase , _lowerCamelCase ) if j > 0: UpperCamelCase_: Tuple = self.blend_h(row[j - 1] , _lowerCamelCase , _lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(_lowerCamelCase , dim=3 ) ) UpperCamelCase_: str = torch.cat(_lowerCamelCase , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=_lowerCamelCase ) def _a ( self , _lowerCamelCase , _lowerCamelCase = False , _lowerCamelCase = True , _lowerCamelCase = None , ): UpperCamelCase_: List[Any] = sample UpperCamelCase_: Optional[Any] = self.encode(_lowerCamelCase ).latent_dist if sample_posterior: UpperCamelCase_: int = posterior.sample(generator=_lowerCamelCase ) else: UpperCamelCase_: Dict = posterior.mode() UpperCamelCase_: Dict = self.decode(_lowerCamelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=_lowerCamelCase )
57
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: str ): def get_matched_characters(lowerCAmelCase_: str , lowerCAmelCase_: str ) -> str: snake_case_ : Tuple = [] snake_case_ : Tuple = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): snake_case_ : str = int(max(0 , i - limit ) ) snake_case_ : Optional[int] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) snake_case_ : List[Any] = f"{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}" return "".join(lowerCAmelCase_ ) # matching characters snake_case_ : List[Any] = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : int = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = len(lowerCAmelCase_ ) # transposition snake_case_ : List[str] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: snake_case_ : str = 0.0 else: snake_case_ : Optional[Any] = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters snake_case_ : Optional[Any] = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("hello", "world"))
666
0
"""simple docstring""" from __future__ import annotations __lowerCAmelCase : Optional[Any] = '''#''' class _lowerCAmelCase : """simple docstring""" def __init__( self ) -> None: '''simple docstring''' snake_case_ : dict = {} def UpperCAmelCase__ ( self , _lowercase ) -> None: '''simple docstring''' snake_case_ : Optional[int] = self._trie for char in text: if char not in trie: snake_case_ : Dict = {} snake_case_ : List[str] = trie[char] snake_case_ : List[Any] = True def UpperCAmelCase__ ( self , _lowercase ) -> tuple | list: '''simple docstring''' snake_case_ : Optional[Any] = self._trie for char in prefix: if char in trie: snake_case_ : Any = trie[char] else: return [] return self._elements(_lowercase ) def UpperCAmelCase__ ( self , _lowercase ) -> tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] for c, v in d.items(): snake_case_ : str = [""" """] if c == END else [(c + s) for s in self._elements(_lowercase )] result.extend(_lowercase ) return tuple(_lowercase ) __lowerCAmelCase : List[str] = Trie() __lowerCAmelCase : List[str] = ('''depart''', '''detergent''', '''daring''', '''dog''', '''deer''', '''deal''') for word in words: trie.insert_word(word) def __lowerCAmelCase ( __UpperCamelCase : str ): '''simple docstring''' snake_case_ : int = trie.find_word(__UpperCamelCase ) return tuple(string + word for word in suffixes ) def __lowerCAmelCase ( ): '''simple docstring''' print(autocomplete_using_trie("""de""" ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) set_seed(7_7_0) UpperCAmelCase = { "c_attn": "att_proj", "c_proj": "out_proj", "c_fc": "in_proj", "transformer.": "", "h.": "layers.", "ln_1": "layernorm_1", "ln_2": "layernorm_2", "ln_f": "layernorm_final", "wpe": "position_embeds_layer", "wte": "input_embeds_layer", } UpperCAmelCase = { "text_small": { "repo_id": "suno/bark", "file_name": "text.pt", }, "coarse_small": { "repo_id": "suno/bark", "file_name": "coarse.pt", }, "fine_small": { "repo_id": "suno/bark", "file_name": "fine.pt", }, "text": { "repo_id": "suno/bark", "file_name": "text_2.pt", }, "coarse": { "repo_id": "suno/bark", "file_name": "coarse_2.pt", }, "fine": { "repo_id": "suno/bark", "file_name": "fine_2.pt", }, } UpperCAmelCase = os.path.dirname(os.path.abspath(__file__)) UpperCAmelCase = os.path.join(os.path.expanduser("~"), ".cache") UpperCAmelCase = os.path.join(os.getenv("XDG_CACHE_HOME", default_cache_dir), "suno", "bark_v0") def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int , lowerCAmelCase_: List[str]=False ): snake_case_ : Union[str, Any] = model_type if use_small: key += "_small" return os.path.join(lowerCAmelCase_ , REMOTE_MODEL_PATHS[key]["file_name"] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: List[str] ): os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) hf_hub_download(repo_id=lowerCAmelCase_ , filename=lowerCAmelCase_ , local_dir=lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Any , lowerCAmelCase_: Dict , lowerCAmelCase_: List[str]=False , lowerCAmelCase_: Dict="text" ): if model_type == "text": snake_case_ : int = BarkSemanticModel snake_case_ : str = BarkSemanticConfig snake_case_ : Optional[Any] = BarkSemanticGenerationConfig elif model_type == "coarse": snake_case_ : str = BarkCoarseModel snake_case_ : Optional[int] = BarkCoarseConfig snake_case_ : Any = BarkCoarseGenerationConfig elif model_type == "fine": snake_case_ : Optional[int] = BarkFineModel snake_case_ : Tuple = BarkFineConfig snake_case_ : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() snake_case_ : Optional[Any] = f"{model_type}_small" if use_small else model_type snake_case_ : Any = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(lowerCAmelCase_ ): logger.info(f"{model_type} model not found, downloading into `{CACHE_DIR}`." ) _download(model_info["repo_id"] , model_info["file_name"] ) snake_case_ : Any = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) # this is a hack snake_case_ : Union[str, Any] = checkpoint["model_args"] if "input_vocab_size" not in model_args: snake_case_ : str = model_args["vocab_size"] snake_case_ : Union[str, Any] = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments snake_case_ : Union[str, Any] = model_args.pop("n_head" ) snake_case_ : int = model_args.pop("n_embd" ) snake_case_ : Any = model_args.pop("n_layer" ) snake_case_ : List[str] = ConfigClass(**checkpoint["model_args"] ) snake_case_ : Optional[Any] = ModelClass(config=lowerCAmelCase_ ) snake_case_ : Tuple = GenerationConfigClass() snake_case_ : List[str] = model_generation_config snake_case_ : Optional[int] = checkpoint["model"] # fixup checkpoint snake_case_ : Optional[int] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(lowerCAmelCase_ ): # replace part of the key with corresponding layer name in HF implementation snake_case_ : Tuple = k[len(lowerCAmelCase_ ) :] for old_layer_name in new_layer_name_dict: snake_case_ : int = new_k.replace(lowerCAmelCase_ , new_layer_name_dict[old_layer_name] ) snake_case_ : int = state_dict.pop(lowerCAmelCase_ ) snake_case_ : Optional[int] = set(state_dict.keys() ) - set(model.state_dict().keys() ) snake_case_ : str = {k for k in extra_keys if not k.endswith(".attn.bias" )} snake_case_ : Any = set(model.state_dict().keys() ) - set(state_dict.keys() ) snake_case_ : List[Any] = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(lowerCAmelCase_ ) != 0: raise ValueError(f"extra keys found: {extra_keys}" ) if len(lowerCAmelCase_ ) != 0: raise ValueError(f"missing keys: {missing_keys}" ) model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) snake_case_ : str = model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) snake_case_ : Union[str, Any] = checkpoint["best_val_loss"].item() logger.info(f"model loaded: {round(n_params/1e6 , 1 )}M params, {round(lowerCAmelCase_ , 3 )} loss" ) model.eval() model.to(lowerCAmelCase_ ) del checkpoint, state_dict return model def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: List[Any] , lowerCAmelCase_: str=False , lowerCAmelCase_: int="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() snake_case_ : int = "cpu" # do conversion on cpu snake_case_ : Optional[Any] = _get_ckpt_path(lowerCAmelCase_ , use_small=lowerCAmelCase_ ) snake_case_ : Tuple = _load_model(lowerCAmelCase_ , lowerCAmelCase_ , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) # load bark initial model snake_case_ : int = _bark_load_model(lowerCAmelCase_ , "cpu" , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) if model_type == "text": snake_case_ : Union[str, Any] = bark_model["model"] if model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model snake_case_ : Optional[Any] = 5 snake_case_ : Optional[int] = 1_0 if model_type in ["text", "coarse"]: snake_case_ : Optional[Any] = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) snake_case_ : str = bark_model(lowerCAmelCase_ )[0] snake_case_ : Tuple = model(lowerCAmelCase_ ) # take last logits snake_case_ : List[str] = output_new_model_total.logits[:, [-1], :] else: snake_case_ : Optional[int] = 3 snake_case_ : str = 8 snake_case_ : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) snake_case_ : Any = model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Union[str, Any] = bark_model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: List[str] , lowerCAmelCase_: Any , lowerCAmelCase_: List[Any] , lowerCAmelCase_: int , lowerCAmelCase_: Optional[Any] , ): snake_case_ : Optional[Any] = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkSemanticConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = BarkCoarseConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[str] = BarkFineConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) snake_case_ : List[str] = BarkSemanticModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkCoarseModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Tuple = BarkFineModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Union[str, Any] = EncodecModel.from_pretrained("facebook/encodec_24khz" ) snake_case_ : Tuple = BarkConfig.from_sub_model_configs( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : List[Any] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) snake_case_ : Optional[int] = BarkModel(lowerCAmelCase_ ) snake_case_ : int = semantic snake_case_ : List[str] = coarseAcoustic snake_case_ : str = fineAcoustic snake_case_ : Optional[Any] = codec snake_case_ : Any = bark_generation_config Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) bark.save_pretrained(lowerCAmelCase_ , repo_id=lowerCAmelCase_ , push_to_hub=lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("model_type", type=str, help="text, coarse or fine.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--is_small", action="store_true", help="convert the small version instead of the large.") UpperCAmelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
666
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig __A = logging.get_logger(__name__) __A = { "Intel/dpt-large": "https://huggingface.co/Intel/dpt-large/resolve/main/config.json", # See all DPT models at https://huggingface.co/models?filter=dpt } class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "dpt" def __init__(self : Optional[int] , UpperCAmelCase_ : Any=768 , UpperCAmelCase_ : Optional[Any]=12 , UpperCAmelCase_ : Union[str, Any]=12 , UpperCAmelCase_ : Tuple=3_072 , UpperCAmelCase_ : Tuple="gelu" , UpperCAmelCase_ : Dict=0.0 , UpperCAmelCase_ : List[str]=0.0 , UpperCAmelCase_ : List[Any]=0.02 , UpperCAmelCase_ : Any=1E-1_2 , UpperCAmelCase_ : Tuple=384 , UpperCAmelCase_ : Dict=16 , UpperCAmelCase_ : str=3 , UpperCAmelCase_ : Any=False , UpperCAmelCase_ : Dict=True , UpperCAmelCase_ : Dict=[2, 5, 8, 11] , UpperCAmelCase_ : List[str]="project" , UpperCAmelCase_ : Dict=[4, 2, 1, 0.5] , UpperCAmelCase_ : Dict=[96, 192, 384, 768] , UpperCAmelCase_ : Dict=256 , UpperCAmelCase_ : List[Any]=-1 , UpperCAmelCase_ : List[str]=False , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : List[str]=0.4 , UpperCAmelCase_ : str=255 , UpperCAmelCase_ : Optional[int]=0.1 , UpperCAmelCase_ : int=[1, 1_024, 24, 24] , UpperCAmelCase_ : Dict=[0, 1] , UpperCAmelCase_ : List[str]=None , **UpperCAmelCase_ : Optional[Any] , ) ->Tuple: '''simple docstring''' super().__init__(**UpperCAmelCase_) lowerCamelCase__: Optional[int] =hidden_size lowerCamelCase__: List[str] =is_hybrid if self.is_hybrid: if backbone_config is None: logger.info("Initializing the config with a `BiT` backbone.") lowerCamelCase__: str ={ "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, } lowerCamelCase__: Union[str, Any] =BitConfig(**UpperCAmelCase_) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_): logger.info("Initializing the config with a `BiT` backbone.") lowerCamelCase__: Optional[Any] =BitConfig(**UpperCAmelCase_) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_): lowerCamelCase__: int =backbone_config else: raise ValueError( F"""backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}.""") lowerCamelCase__: Optional[int] =backbone_featmap_shape lowerCamelCase__: Optional[Any] =neck_ignore_stages if readout_type != "project": raise ValueError("Readout type must be 'project' when using `DPT-hybrid` mode.") else: lowerCamelCase__: List[Any] =None lowerCamelCase__: Optional[int] =None lowerCamelCase__: Optional[Any] =[] lowerCamelCase__: Union[str, Any] =num_hidden_layers lowerCamelCase__: List[str] =num_attention_heads lowerCamelCase__: Any =intermediate_size lowerCamelCase__: Dict =hidden_act lowerCamelCase__: Optional[int] =hidden_dropout_prob lowerCamelCase__: Dict =attention_probs_dropout_prob lowerCamelCase__: Any =initializer_range lowerCamelCase__: Optional[int] =layer_norm_eps lowerCamelCase__: Dict =image_size lowerCamelCase__: str =patch_size lowerCamelCase__: Any =num_channels lowerCamelCase__: Dict =qkv_bias lowerCamelCase__: Dict =backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError("Readout_type must be one of ['ignore', 'add', 'project']") lowerCamelCase__: int =readout_type lowerCamelCase__: str =reassemble_factors lowerCamelCase__: int =neck_hidden_sizes lowerCamelCase__: Dict =fusion_hidden_size lowerCamelCase__: List[str] =head_in_index lowerCamelCase__: Dict =use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) lowerCamelCase__: Dict =use_auxiliary_head lowerCamelCase__: List[str] =auxiliary_loss_weight lowerCamelCase__: Union[str, Any] =semantic_loss_ignore_index lowerCamelCase__: int =semantic_classifier_dropout def SCREAMING_SNAKE_CASE_ (self : Dict) ->Tuple: '''simple docstring''' lowerCamelCase__: List[str] =copy.deepcopy(self.__dict__) if output["backbone_config"] is not None: lowerCamelCase__: List[Any] =self.backbone_config.to_dict() lowerCamelCase__: Union[str, Any] =self.__class__.model_type return output
59
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ = { '''configuration_albert''': ['''ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AlbertConfig''', '''AlbertOnnxConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''AlbertTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''AlbertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AlbertForMaskedLM''', '''AlbertForMultipleChoice''', '''AlbertForPreTraining''', '''AlbertForQuestionAnswering''', '''AlbertForSequenceClassification''', '''AlbertForTokenClassification''', '''AlbertModel''', '''AlbertPreTrainedModel''', '''load_tf_weights_in_albert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''TF_ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFAlbertForMaskedLM''', '''TFAlbertForMultipleChoice''', '''TFAlbertForPreTraining''', '''TFAlbertForQuestionAnswering''', '''TFAlbertForSequenceClassification''', '''TFAlbertForTokenClassification''', '''TFAlbertMainLayer''', '''TFAlbertModel''', '''TFAlbertPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''FlaxAlbertForMaskedLM''', '''FlaxAlbertForMultipleChoice''', '''FlaxAlbertForPreTraining''', '''FlaxAlbertForQuestionAnswering''', '''FlaxAlbertForSequenceClassification''', '''FlaxAlbertForTokenClassification''', '''FlaxAlbertModel''', '''FlaxAlbertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_albert import ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, AlbertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_albert import AlbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_albert_fast import AlbertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_albert import ( ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, AlbertPreTrainedModel, load_tf_weights_in_albert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_albert import ( TF_ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFAlbertForMaskedLM, TFAlbertForMultipleChoice, TFAlbertForPreTraining, TFAlbertForQuestionAnswering, TFAlbertForSequenceClassification, TFAlbertForTokenClassification, TFAlbertMainLayer, TFAlbertModel, TFAlbertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, FlaxAlbertPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
60
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch UpperCAmelCase = logging.get_logger(__name__) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : str = ["pixel_values"] def __init__( self : List[Any] , A__ : bool = True , A__ : Optional[Dict[str, int]] = None , A__ : PILImageResampling = PILImageResampling.BILINEAR , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , **A__ : int , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[int] = size if size is not None else {"shortest_edge": 2_56} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[str] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Any = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : int = do_resize snake_case_ : Optional[Any] = size snake_case_ : Optional[Any] = resample snake_case_ : Optional[int] = do_center_crop snake_case_ : List[Any] = crop_size snake_case_ : List[Any] = do_rescale snake_case_ : Optional[int] = rescale_factor snake_case_ : Optional[Any] = do_normalize snake_case_ : List[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : str , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : Any = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : int , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Union[str, Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Tuple = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : float , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Tuple ) -> np.ndarray: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Dict , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , A__ : ImageInput , A__ : Optional[bool] = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : Dict[str, int] = None , A__ : Optional[bool] = None , A__ : Optional[float] = None , A__ : Optional[bool] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **A__ : Union[str, Any] , ) -> Optional[int]: '''simple docstring''' snake_case_ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Dict = size if size is not None else self.size snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : str = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : Dict = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Any = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : Dict = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Tuple = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Any = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : List[str] = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : Any = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : Union[str, Any] = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : Optional[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Any = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : Dict , A__ : List[Tuple] = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A__ ) != len(A__ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(A__ ): snake_case_ : Dict = target_sizes.numpy() snake_case_ : int = [] for idx in range(len(A__ ) ): snake_case_ : List[str] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=A__ ) snake_case_ : int = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A__ ) else: snake_case_ : List[Any] = logits.argmax(dim=1 ) snake_case_ : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
666
0
from ....configuration_utils import PretrainedConfig from ....utils import logging UpperCamelCase = logging.get_logger(__name__) # TODO: upload to AWS UpperCamelCase = { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json' ), } class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = "retribert" def __init__( self : int , SCREAMING_SNAKE_CASE__ : Union[str, Any]=30_522 , SCREAMING_SNAKE_CASE__ : Optional[Any]=768 , SCREAMING_SNAKE_CASE__ : Optional[int]=8 , SCREAMING_SNAKE_CASE__ : List[Any]=12 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3_072 , SCREAMING_SNAKE_CASE__ : int="gelu" , SCREAMING_SNAKE_CASE__ : Optional[int]=0.1 , SCREAMING_SNAKE_CASE__ : Optional[int]=0.1 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=512 , SCREAMING_SNAKE_CASE__ : Optional[Any]=2 , SCREAMING_SNAKE_CASE__ : Tuple=0.02 , SCREAMING_SNAKE_CASE__ : List[Any]=1e-1_2 , SCREAMING_SNAKE_CASE__ : str=True , SCREAMING_SNAKE_CASE__ : str=128 , SCREAMING_SNAKE_CASE__ : str=0 , **SCREAMING_SNAKE_CASE__ : List[Any] , ) -> Tuple: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = hidden_act lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = type_vocab_size lowerCAmelCase__ = initializer_range lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = share_encoders lowerCAmelCase__ = projection_dim
61
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class SCREAMING_SNAKE_CASE : '''simple docstring''' @staticmethod def _A ( *UpperCAmelCase_ : Optional[Any] , **UpperCAmelCase_ : Union[str, Any] ): pass def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' UpperCamelCase_ : int = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def _A ( self : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict ): SCREAMING_SNAKE_CASE : Tuple = DepthEstimationPipeline(model=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def _A ( self : Optional[int] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] ): SCREAMING_SNAKE_CASE : List[str] = depth_estimator("./tests/fixtures/tests_samples/COCO/000000039769.png" ) self.assertEqual({"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )} , UpperCAmelCase_ ) import datasets SCREAMING_SNAKE_CASE : int = datasets.load_dataset("hf-internal-testing/fixtures_image_utils" , "image" , split="test" ) SCREAMING_SNAKE_CASE : List[Any] = depth_estimator( [ Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ), "http://images.cocodataset.org/val2017/000000039769.jpg", # RGBA dataset[0]["file"], # LA dataset[1]["file"], # L dataset[2]["file"], ] ) self.assertEqual( [ {"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )}, {"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )}, {"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )}, {"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )}, {"predicted_depth": ANY(torch.Tensor ), "depth": ANY(Image.Image )}, ] , UpperCAmelCase_ , ) @require_tf @unittest.skip("Depth estimation is not implemented in TF" ) def _A ( self : Tuple ): pass @slow @require_torch def _A ( self : int ): SCREAMING_SNAKE_CASE : List[Any] = "Intel/dpt-large" SCREAMING_SNAKE_CASE : Optional[Any] = pipeline("depth-estimation" , model=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = depth_estimator("http://images.cocodataset.org/val2017/000000039769.jpg" ) SCREAMING_SNAKE_CASE : Optional[int] = hashimage(outputs["depth"] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs["predicted_depth"].max().item() ) , 29.304 ) self.assertEqual(nested_simplify(outputs["predicted_depth"].min().item() ) , 2.662 ) @require_torch def _A ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest("There is not hf-internal-testing tiny model for either GLPN nor DPT" )
62
from ...configuration_utils import PretrainedConfig UpperCAmelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "tapas" def __init__( self : List[Any] , A__ : str=3_05_22 , A__ : Tuple=7_68 , A__ : List[Any]=12 , A__ : Optional[Any]=12 , A__ : Union[str, Any]=30_72 , A__ : Dict="gelu" , A__ : List[Any]=0.1 , A__ : str=0.1 , A__ : List[Any]=10_24 , A__ : Optional[int]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , A__ : Union[str, Any]=0.02 , A__ : Tuple=1E-12 , A__ : Tuple=0 , A__ : Any=10.0 , A__ : List[str]=0 , A__ : List[str]=1.0 , A__ : Optional[Any]=None , A__ : Tuple=1.0 , A__ : Union[str, Any]=False , A__ : Any=None , A__ : Union[str, Any]=1.0 , A__ : int=1.0 , A__ : str=False , A__ : int=False , A__ : Optional[Any]="ratio" , A__ : str=None , A__ : int=None , A__ : Dict=64 , A__ : int=32 , A__ : Optional[Any]=False , A__ : List[str]=True , A__ : List[Any]=False , A__ : str=False , A__ : Any=True , A__ : Tuple=False , A__ : str=None , A__ : str=None , **A__ : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=A__ , **A__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) snake_case_ : int = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[int] = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : List[Any] = type_vocab_sizes snake_case_ : str = initializer_range snake_case_ : Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters snake_case_ : Optional[int] = positive_label_weight snake_case_ : Dict = num_aggregation_labels snake_case_ : List[str] = aggregation_loss_weight snake_case_ : str = use_answer_as_supervision snake_case_ : int = answer_loss_importance snake_case_ : Any = use_normalized_answer_loss snake_case_ : int = huber_loss_delta snake_case_ : List[Any] = temperature snake_case_ : str = aggregation_temperature snake_case_ : List[str] = use_gumbel_for_cells snake_case_ : List[str] = use_gumbel_for_aggregation snake_case_ : Dict = average_approximation_function snake_case_ : List[str] = cell_selection_preference snake_case_ : Dict = answer_loss_cutoff snake_case_ : List[str] = max_num_rows snake_case_ : Union[str, Any] = max_num_columns snake_case_ : str = average_logits_per_cell snake_case_ : Union[str, Any] = select_one_column snake_case_ : Dict = allow_empty_column_selection snake_case_ : List[Any] = init_cell_selection_weights_to_zero snake_case_ : str = reset_position_index_per_cell snake_case_ : List[Any] = disable_per_token_loss # Aggregation hyperparameters snake_case_ : List[str] = aggregation_labels snake_case_ : Union[str, Any] = no_aggregation_label_index if isinstance(self.aggregation_labels , A__ ): snake_case_ : Optional[int] = {int(A__ ): v for k, v in aggregation_labels.items()}
666
0
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class a ( lowercase__ ): """simple docstring""" a : jnp.ndarray @flax_register_to_config class a ( nn.Module , lowercase__ , lowercase__ ): """simple docstring""" a : int = 32 a : int = 4 a : int = 4 a : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) a : Tuple[str] = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") a : Union[bool, Tuple[bool]] = False a : Tuple[int] = (320, 640, 1_280, 1_280) a : int = 2 a : Union[int, Tuple[int]] = 8 a : Optional[Union[int, Tuple[int]]] = None a : int = 1_280 a : float = 0.0 a : bool = False a : jnp.dtype = jnp.floataa a : bool = True a : int = 0 a : bool = False def UpperCAmelCase ( self : int , __lowercase : jax.random.KeyArray ) -> FrozenDict: # init input tensors __UpperCAmelCase : List[str] = (1, self.in_channels, self.sample_size, self.sample_size) __UpperCAmelCase : Optional[int] = jnp.zeros(__lowercase , dtype=jnp.floataa ) __UpperCAmelCase : Optional[int] = jnp.ones((1,) , dtype=jnp.intaa ) __UpperCAmelCase : Optional[int] = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) __UpperCAmelCase , __UpperCAmelCase : Tuple = jax.random.split(__lowercase ) __UpperCAmelCase : int = {"""params""": params_rng, """dropout""": dropout_rng} return self.init(__lowercase , __lowercase , __lowercase , __lowercase )["params"] def UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: __UpperCAmelCase : str = self.block_out_channels __UpperCAmelCase : Union[str, Any] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( """At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.""" ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. __UpperCAmelCase : Tuple = self.num_attention_heads or self.attention_head_dim # input __UpperCAmelCase : str = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time __UpperCAmelCase : Dict = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) __UpperCAmelCase : Optional[int] = FlaxTimestepEmbedding(__lowercase , dtype=self.dtype ) __UpperCAmelCase : int = self.only_cross_attention if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : List[str] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : List[str] = (num_attention_heads,) * len(self.down_block_types ) # down __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : int = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): __UpperCAmelCase : Dict = output_channel __UpperCAmelCase : Optional[Any] = block_out_channels[i] __UpperCAmelCase : Union[str, Any] = i == len(__lowercase ) - 1 if down_block_type == "CrossAttnDownBlock2D": __UpperCAmelCase : List[str] = FlaxCrossAttnDownBlockaD( in_channels=__lowercase , out_channels=__lowercase , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: __UpperCAmelCase : List[Any] = FlaxDownBlockaD( in_channels=__lowercase , out_channels=__lowercase , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(__lowercase ) __UpperCAmelCase : Union[str, Any] = down_blocks # mid __UpperCAmelCase : Any = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up __UpperCAmelCase : Tuple = [] __UpperCAmelCase : Tuple = list(reversed(__lowercase ) ) __UpperCAmelCase : Any = list(reversed(__lowercase ) ) __UpperCAmelCase : Tuple = list(reversed(__lowercase ) ) __UpperCAmelCase : Tuple = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): __UpperCAmelCase : Dict = output_channel __UpperCAmelCase : Any = reversed_block_out_channels[i] __UpperCAmelCase : Union[str, Any] = reversed_block_out_channels[min(i + 1 , len(__lowercase ) - 1 )] __UpperCAmelCase : Union[str, Any] = i == len(__lowercase ) - 1 if up_block_type == "CrossAttnUpBlock2D": __UpperCAmelCase : Optional[Any] = FlaxCrossAttnUpBlockaD( in_channels=__lowercase , out_channels=__lowercase , prev_output_channel=__lowercase , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: __UpperCAmelCase : Any = FlaxUpBlockaD( in_channels=__lowercase , out_channels=__lowercase , prev_output_channel=__lowercase , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(__lowercase ) __UpperCAmelCase : List[str] = output_channel __UpperCAmelCase : Any = up_blocks # out __UpperCAmelCase : str = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) __UpperCAmelCase : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Optional[int] , __lowercase : List[str] , __lowercase : int , __lowercase : Dict , __lowercase : int=None , __lowercase : Optional[int]=None , __lowercase : bool = True , __lowercase : bool = False , ) -> Union[FlaxUNetaDConditionOutput, Tuple]: # 1. time if not isinstance(__lowercase , jnp.ndarray ): __UpperCAmelCase : Tuple = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(__lowercase , jnp.ndarray ) and len(timesteps.shape ) == 0: __UpperCAmelCase : int = timesteps.astype(dtype=jnp.floataa ) __UpperCAmelCase : List[Any] = jnp.expand_dims(__lowercase , 0 ) __UpperCAmelCase : List[str] = self.time_proj(__lowercase ) __UpperCAmelCase : Tuple = self.time_embedding(__lowercase ) # 2. pre-process __UpperCAmelCase : Any = jnp.transpose(__lowercase , (0, 2, 3, 1) ) __UpperCAmelCase : Optional[int] = self.conv_in(__lowercase ) # 3. down __UpperCAmelCase : str = (sample,) for down_block in self.down_blocks: if isinstance(__lowercase , __lowercase ): __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = down_block(__lowercase , __lowercase , __lowercase , deterministic=not train ) else: __UpperCAmelCase , __UpperCAmelCase : int = down_block(__lowercase , __lowercase , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: __UpperCAmelCase : str = () for down_block_res_sample, down_block_additional_residual in zip( __lowercase , __lowercase ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) __UpperCAmelCase : Tuple = new_down_block_res_samples # 4. mid __UpperCAmelCase : List[str] = self.mid_block(__lowercase , __lowercase , __lowercase , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: __UpperCAmelCase : Optional[Any] = down_block_res_samples[-(self.layers_per_block + 1) :] __UpperCAmelCase : Optional[int] = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(__lowercase , __lowercase ): __UpperCAmelCase : List[Any] = up_block( __lowercase , temb=__lowercase , encoder_hidden_states=__lowercase , res_hidden_states_tuple=__lowercase , deterministic=not train , ) else: __UpperCAmelCase : Union[str, Any] = up_block(__lowercase , temb=__lowercase , res_hidden_states_tuple=__lowercase , deterministic=not train ) # 6. post-process __UpperCAmelCase : Any = self.conv_norm_out(__lowercase ) __UpperCAmelCase : Tuple = nn.silu(__lowercase ) __UpperCAmelCase : List[str] = self.conv_out(__lowercase ) __UpperCAmelCase : str = jnp.transpose(__lowercase , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=__lowercase )
63
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : str , A__ : str ) -> Optional[int]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : int , A__ : Optional[int] , A__ : Dict ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Any , A__ : List[str] , A__ : str ) -> Optional[int]: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : List[Any] , A__ : List[str] , A__ : Optional[int] ) -> List[str]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class snake_case__ ( _UpperCamelCase ): @require_beam def UpperCAmelCase__ ( self : str ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Dict = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[int] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : str ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam snake_case_ : Tuple = beam.io.parquetio.WriteToParquet snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[Any] = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: snake_case_ : int = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[Any] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Tuple = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[str] = NestedBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) snake_case_ : int = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
666
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
64
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple ): class snake_case__ : def __init__( self : Any , A__ : List[str] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = metric_id class snake_case__ : _SCREAMING_SNAKE_CASE : List[str] = [MetricMock(_UpperCamelCase ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: int , lowerCAmelCase_: List[Any] , lowerCAmelCase_: Any , lowerCAmelCase_: List[str] ): if "tmp_path" in args: snake_case_ : List[Any] = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowerCAmelCase_ , match="https://huggingface.co/docs/evaluate" ): func(*lowerCAmelCase_ )
666
0
"""simple docstring""" from collections.abc import Callable class __lowercase : def __init__( self : Tuple ,A : Callable | None = None ): '''simple docstring''' # Stores actual heap items. UpperCAmelCase__ : list = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase__ : dict = {} # Stores current size of heap. UpperCAmelCase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase__ : int = key or (lambda A : x) def __lowercase ( self : Union[str, Any] ,A : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def __lowercase ( self : Tuple ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = int(2 * i + 1 ) return left if 0 < left < self.size else None def __lowercase ( self : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = int(2 * i + 2 ) return right if 0 < right < self.size else None def __lowercase ( self : List[Any] ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : int = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.arr[j], self.arr[i] def __lowercase ( self : Optional[int] ,A : int ,A : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def __lowercase ( self : Optional[int] ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._left(A ) UpperCAmelCase__ : Dict = self._right(A ) UpperCAmelCase__ : Optional[int] = i if left is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = left if right is not None and not self._cmp(A ,A ): UpperCAmelCase__ : List[Any] = right return valid_parent def __lowercase ( self : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = self._parent(A ) while parent is not None and not self._cmp(A ,A ): self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : int = parent, self._parent(A ) def __lowercase ( self : str ,A : int ): '''simple docstring''' UpperCAmelCase__ : Any = self._get_valid_parent(A ) while valid_parent != index: self._swap(A ,A ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = valid_parent, self._get_valid_parent(A ) def __lowercase ( self : Optional[Any] ,A : int ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Tuple = self.pos_map[item] UpperCAmelCase__ : Dict = [item, self.key(A )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : List[Any] ,A : int ): '''simple docstring''' if item not in self.pos_map: return UpperCAmelCase__ : Any = self.pos_map[item] del self.pos_map[item] UpperCAmelCase__ : Dict = self.arr[self.size - 1] UpperCAmelCase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(A ) self._heapify_down(A ) def __lowercase ( self : str ,A : int ,A : int ): '''simple docstring''' UpperCAmelCase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(A )] ) else: UpperCAmelCase__ : List[str] = [item, self.key(A )] UpperCAmelCase__ : Union[str, Any] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def __lowercase ( self : str ): '''simple docstring''' return self.arr[0] if self.size else None def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Tuple = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def lowerCAmelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
65
from __future__ import annotations import bisect def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Any = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: snake_case_ : Tuple = mid + 1 else: snake_case_ : Dict = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Optional[Any] = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: snake_case_ : Optional[Any] = mid + 1 else: snake_case_ : Tuple = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_left(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_right(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Dict = 0 snake_case_ : Tuple = len(lowerCAmelCase_ ) - 1 while left <= right: snake_case_ : int = left + (right - left) // 2 snake_case_ : Optional[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: snake_case_ : Optional[Any] = midpoint - 1 else: snake_case_ : Optional[int] = midpoint + 1 return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Optional[int] = bisect.bisect_left(lowerCAmelCase_ , lowerCAmelCase_ ) if index != len(lowerCAmelCase_ ) and sorted_collection[index] == item: return index return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int , lowerCAmelCase_: int ): if right < left: return None snake_case_ : List[Any] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , midpoint - 1 ) else: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , midpoint + 1 , lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = input("Enter numbers separated by comma:\n").strip() UpperCAmelCase = sorted(int(item) for item in user_input.split(",")) UpperCAmelCase = int(input("Enter a single number to be found in the list:\n")) UpperCAmelCase = binary_search(collection, target) if result is None: print(F"{target} was not found in {collection}.") else: print(F"{target} was found at position {result} in {collection}.")
666
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , ) -> Tuple: _lowercase : List[str] = {} if train_file is not None: _lowercase : Tuple = [train_file] if eval_file is not None: _lowercase : Optional[Any] = [eval_file] if test_file is not None: _lowercase : Dict = [test_file] _lowercase : Dict = datasets.load_dataset('csv' , data_files=SCREAMING_SNAKE_CASE ) _lowercase : Dict = list(ds[list(files.keys() )[0]].features.keys() ) _lowercase : List[Any] = features_name.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = list(set(ds[list(files.keys() )[0]][label_name] ) ) _lowercase : Dict = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE )} _lowercase : Tuple = tokenizer.model_input_names _lowercase : int = {} if len(SCREAMING_SNAKE_CASE ) == 1: for k in files.keys(): _lowercase : List[str] = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='max_length' ) , batched=SCREAMING_SNAKE_CASE , ) elif len(SCREAMING_SNAKE_CASE ) == 2: for k in files.keys(): _lowercase : Tuple = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='max_length' , ) , batched=SCREAMING_SNAKE_CASE , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _lowercase : str = {k: v for k, v in ex.items() if k in input_names} _lowercase : List[str] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _lowercase : str = {k: v for k, v in ex.items() if k in input_names} _lowercase : Tuple = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _lowercase : int = {k: v for k, v in ex.items() if k in input_names} _lowercase : str = labelaid[ex[label_name]] yield (d, label) _lowercase : str = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _lowercase : Optional[int] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) _lowercase : int = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _lowercase : str = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) _lowercase : List[str] = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _lowercase : Optional[Any] = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCAmelCase_ : _UpperCamelCase : int = field(metadata={"help": "Which column contains the label"} ) _UpperCamelCase : str = field(default=__snake_case , metadata={"help": "The path of the training file"} ) _UpperCamelCase : Optional[str] = field(default=__snake_case , metadata={"help": "The path of the development file"} ) _UpperCamelCase : Optional[str] = field(default=__snake_case , metadata={"help": "The path of the test file"} ) _UpperCamelCase : int = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _UpperCamelCase : bool = field( default=__snake_case , metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class lowerCAmelCase_ : _UpperCamelCase : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) _UpperCamelCase : bool = field(default=__snake_case , metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) def __magic_name__ ( ) -> str: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) _lowercase , _lowercase , _lowercase : Optional[int] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO , ) logger.info( F"""n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, """ F"""16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _lowercase , _lowercase , _lowercase , _lowercase : int = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=SCREAMING_SNAKE_CASE , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(SCREAMING_SNAKE_CASE ) , labelaid=SCREAMING_SNAKE_CASE , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='text-classification' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _lowercase : List[str] = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('.bin' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) def compute_metrics(SCREAMING_SNAKE_CASE ) -> Dict: _lowercase : Optional[int] = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _lowercase : Optional[Any] = TFTrainer( model=SCREAMING_SNAKE_CASE , args=SCREAMING_SNAKE_CASE , train_dataset=SCREAMING_SNAKE_CASE , eval_dataset=SCREAMING_SNAKE_CASE , compute_metrics=SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation _lowercase : Dict = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : int = trainer.evaluate() _lowercase : List[Any] = os.path.join(training_args.output_dir , 'eval_results.txt' ) with open(SCREAMING_SNAKE_CASE , 'w' ) as writer: logger.info('***** Eval results *****' ) for key, value in result.items(): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) results.update(SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": main()
66
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , A__ : VQModel , A__ : UNetaDModel , A__ : DDIMScheduler ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , A__ : int = 1 , A__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A__ : float = 0.0 , A__ : int = 50 , A__ : Optional[str] = "pil" , A__ : bool = True , **A__ : Optional[Any] , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' snake_case_ : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) snake_case_ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler snake_case_ : Any = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature snake_case_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case_ : List[Any] = {} if accepts_eta: snake_case_ : int = eta for t in self.progress_bar(self.scheduler.timesteps ): snake_case_ : Union[str, Any] = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual snake_case_ : Dict = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case_ : Union[str, Any] = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE snake_case_ : int = self.vqvae.decode(A__ ).sample snake_case_ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ : Optional[int] = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
666
0
def SCREAMING_SNAKE_CASE__ ( snake_case__ :int ) -> "list[int]": if upper_limit < 0: raise ValueError('Limit for the Catalan sequence must be ≥ 0' ) _lowercase = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 _lowercase = 1 if upper_limit > 0: _lowercase = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(snake_case__ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: snake_case = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F"""The Catalan numbers from 0 through {N} are:""") print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
67
from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) snake_case_ : List[str] = precision snake_case_ : Union[str, Any] = ceil(precision / 1_4 ) snake_case_ : List[str] = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : str = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : str = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): snake_case_ : Tuple = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCAmelCase = 5_0 print(F"The first {n} digits of pi is: {pi(n)}")
666
0
from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING __A = logging.get_logger(__name__) @add_end_docstrings(UpperCamelCase ) class _A ( UpperCamelCase ): """simple docstring""" def __init__( self : Optional[Any] , *__SCREAMING_SNAKE_CASE : str , **__SCREAMING_SNAKE_CASE : List[str] ) -> Union[str, Any]: super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) requires_backends(self , """vision""" ) self.check_model_type(__SCREAMING_SNAKE_CASE ) def __call__( self : List[str] , __SCREAMING_SNAKE_CASE : Union[str, List[str], "Image.Image", List["Image.Image"]] , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Tuple: return super().__call__(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def _a ( self : Dict , **__SCREAMING_SNAKE_CASE : int ) -> Any: return {}, {}, {} def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] ) -> int: __UpperCAmelCase =load_image(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =image.size __UpperCAmelCase =self.image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors=self.framework ) return model_inputs def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : List[str] ) -> List[str]: __UpperCAmelCase =self.model(**__SCREAMING_SNAKE_CASE ) return model_outputs def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : List[Any] ) -> Tuple: __UpperCAmelCase =model_outputs.predicted_depth __UpperCAmelCase =torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode="""bicubic""" , align_corners=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =prediction.squeeze().cpu().numpy() __UpperCAmelCase =(output * 255 / np.max(__SCREAMING_SNAKE_CASE )).astype("""uint8""" ) __UpperCAmelCase =Image.fromarray(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={} __UpperCAmelCase =predicted_depth __UpperCAmelCase =depth return output_dict
68
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int = 1_0_0_0 ): snake_case_ ,snake_case_ : List[str] = 1, 1 snake_case_ : List[str] = 2 while True: snake_case_ : Tuple = 0 snake_case_ : Union[str, Any] = fa + fa snake_case_ ,snake_case_ : str = fa, f index += 1 for _ in str(lowerCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
666
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int | float] , lowerCAmelCase_: int , lowerCAmelCase_: int ): if len(lowerCAmelCase_ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(lowerCAmelCase_ ) or left < -len(lowerCAmelCase_ ) or right >= len(lowerCAmelCase_ ) or right < -len(lowerCAmelCase_ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case_ : List[Any] = (left + right) >> 1 # the middle snake_case_ : Dict = find_max(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # find max in range[left, mid] snake_case_ : int = find_max(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
666
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase : Dict = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[str] = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[str] = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowerCamelCase : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure)
70
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } UpperCAmelCase = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE : List[str] = RobertaTokenizer def __init__( self : Optional[int] , A__ : List[Any]=None , A__ : Optional[int]=None , A__ : List[str]=None , A__ : Dict="replace" , A__ : List[str]="<s>" , A__ : Optional[Any]="</s>" , A__ : List[str]="</s>" , A__ : List[Any]="<s>" , A__ : int="<unk>" , A__ : int="<pad>" , A__ : List[Any]="<mask>" , A__ : Any=False , A__ : Optional[int]=True , **A__ : Union[str, Any] , ) -> int: '''simple docstring''' super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : List[Any] = getattr(A__ , pre_tok_state.pop("type" ) ) snake_case_ : Any = add_prefix_space snake_case_ : List[Any] = pre_tok_class(**A__ ) snake_case_ : Optional[int] = add_prefix_space snake_case_ : List[str] = "post_processor" snake_case_ : Tuple = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case_ : str = tuple(state["sep"] ) if "cls" in state: snake_case_ : Tuple = tuple(state["cls"] ) snake_case_ : Tuple = False if state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : Optional[Any] = add_prefix_space snake_case_ : str = True if state.get("trim_offsets" , A__ ) != trim_offsets: snake_case_ : Optional[int] = trim_offsets snake_case_ : List[Any] = True if changes_to_apply: snake_case_ : int = getattr(A__ , state.pop("type" ) ) snake_case_ : List[Any] = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCAmelCase__ ( self : Optional[Any] ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCAmelCase__ ( self : Tuple , A__ : Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case_ : Any = value def UpperCAmelCase__ ( self : int , *A__ : Optional[Any] , **A__ : int ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[Any] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , *A__ : Any , **A__ : List[Any] ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[int] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCAmelCase__ ( self : int , A__ : List[str] , A__ : Union[str, Any]=None ) -> Any: '''simple docstring''' snake_case_ : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Dict , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : str = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
666
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _snake_case (__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase): __A : Any =AltDiffusionPipeline __A : List[str] =TEXT_TO_IMAGE_PARAMS __A : Optional[int] =TEXT_TO_IMAGE_BATCH_PARAMS __A : List[Any] =TEXT_TO_IMAGE_IMAGE_PARAMS __A : Dict =TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) UpperCAmelCase_ : int = DDIMScheduler( beta_start=0.00085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=_snake_case ,set_alpha_to_one=_snake_case ,) torch.manual_seed(0 ) UpperCAmelCase_ : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase_ : str = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,projection_dim=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=50_02 ,) UpperCAmelCase_ : Union[str, Any] = CLIPTextModel(_snake_case ) UpperCAmelCase_ : Optional[Any] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) UpperCAmelCase_ : List[Any] = 77 UpperCAmelCase_ : Optional[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase__ ( self ,_snake_case ,_snake_case=0 ): if str(_snake_case ).startswith("mps" ): UpperCAmelCase_ : int = torch.manual_seed(_snake_case ) else: UpperCAmelCase_ : Dict = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[Any] = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase_ : Union[str, Any] = RobertaSeriesConfig( hidden_size=32 ,project_dim=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,vocab_size=50_02 ,) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ : Any = RobertaSeriesModelWithTransformation(_snake_case ) UpperCAmelCase_ : int = text_encoder UpperCAmelCase_ : int = AltDiffusionPipeline(**_snake_case ) UpperCAmelCase_ : Dict = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) UpperCAmelCase_ : Any = self.get_dummy_inputs(_snake_case ) UpperCAmelCase_ : Any = "A photo of an astronaut" UpperCAmelCase_ : Tuple = alt_pipe(**_snake_case ) UpperCAmelCase_ : Union[str, Any] = output.images UpperCAmelCase_ : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : int = np.array( [0.5748162, 0.60447145, 0.48821217, 0.50100636, 0.5431185, 0.45763683, 0.49657696, 0.48132733, 0.47573093] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[int] = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : int = self.get_dummy_components() UpperCAmelCase_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=_snake_case ) torch.manual_seed(0 ) UpperCAmelCase_ : List[str] = RobertaSeriesConfig( hidden_size=32 ,project_dim=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,vocab_size=50_02 ,) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ : Union[str, Any] = RobertaSeriesModelWithTransformation(_snake_case ) UpperCAmelCase_ : Tuple = text_encoder UpperCAmelCase_ : List[str] = AltDiffusionPipeline(**_snake_case ) UpperCAmelCase_ : int = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) UpperCAmelCase_ : Dict = self.get_dummy_inputs(_snake_case ) UpperCAmelCase_ : List[Any] = alt_pipe(**_snake_case ) UpperCAmelCase_ : Optional[Any] = output.images UpperCAmelCase_ : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : Optional[int] = np.array( [0.51605093, 0.5707241, 0.47365507, 0.50578886, 0.5633877, 0.4642503, 0.5182081, 0.48763484, 0.49084237] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class _snake_case (unittest.TestCase): def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): # make sure here that pndm scheduler skips prk UpperCAmelCase_ : Optional[Any] = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" ,safety_checker=_snake_case ) UpperCAmelCase_ : str = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) UpperCAmelCase_ : str = "A painting of a squirrel eating a burger" UpperCAmelCase_ : Any = torch.manual_seed(0 ) UpperCAmelCase_ : Optional[Any] = alt_pipe([prompt] ,generator=_snake_case ,guidance_scale=6.0 ,num_inference_steps=20 ,output_type="np" ) UpperCAmelCase_ : str = output.images UpperCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCAmelCase_ : Union[str, Any] = np.array([0.1010, 0.0800, 0.0794, 0.0885, 0.0843, 0.0762, 0.0769, 0.0729, 0.0586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[str] = DDIMScheduler.from_pretrained("BAAI/AltDiffusion" ,subfolder="scheduler" ) UpperCAmelCase_ : Union[str, Any] = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" ,scheduler=_snake_case ,safety_checker=_snake_case ) UpperCAmelCase_ : int = alt_pipe.to(_snake_case ) alt_pipe.set_progress_bar_config(disable=_snake_case ) UpperCAmelCase_ : List[Any] = "A painting of a squirrel eating a burger" UpperCAmelCase_ : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase_ : Dict = alt_pipe([prompt] ,generator=_snake_case ,num_inference_steps=2 ,output_type="numpy" ) UpperCAmelCase_ : int = output.images UpperCAmelCase_ : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCAmelCase_ : Union[str, Any] = np.array([0.4019, 0.4052, 0.3810, 0.4119, 0.3916, 0.3982, 0.4651, 0.4195, 0.5323] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
71
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) snake_case_ : Any = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } snake_case_ : List[str] = model(A__ )["last_hidden_state"] snake_case_ : str = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , A__ ) # compare the actual values for a slice. snake_case_ : List[str] = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
666
0
'''simple docstring''' import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any]=1_0 ) -> List[Any]: '''simple docstring''' lowercase =[] for _ in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : str=1_0 ) -> Dict: '''simple docstring''' lowercase =[] for step in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: lowercase =os.path.join(lowercase_ , '''schedule.bin''' ) torch.save(scheduler.state_dict() , lowercase_ ) lowercase =torch.load(lowercase_ ) scheduler.load_state_dict(lowercase_ ) return lrs @require_torch class __magic_name__ ( unittest.TestCase ): def _A( self , snake_case_ , snake_case_ , snake_case_ ): self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for a, b in zip(snake_case_ , snake_case_ ): self.assertAlmostEqual(snake_case_ , snake_case_ , delta=snake_case_ ) def _A( self ): lowercase =torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case_ ) lowercase =torch.tensor([0.4, 0.2, -0.5] ) lowercase =nn.MSELoss() # No warmup, constant schedule, no gradient clipping lowercase =AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(1_00 ): lowercase =criterion(snake_case_ , snake_case_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def _A( self ): lowercase =torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case_ ) lowercase =torch.tensor([0.4, 0.2, -0.5] ) lowercase =nn.MSELoss() # No warmup, constant schedule, no gradient clipping lowercase =Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=snake_case_ , weight_decay=0.0 , relative_step=snake_case_ , scale_parameter=snake_case_ , warmup_init=snake_case_ , ) for _ in range(10_00 ): lowercase =criterion(snake_case_ , snake_case_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class __magic_name__ ( unittest.TestCase ): UpperCamelCase__ = nn.Linear(50 , 50 ) if is_torch_available() else None UpperCamelCase__ = AdamW(m.parameters() , lr=1_0.0 ) if is_torch_available() else None UpperCamelCase__ = 10 def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for a, b in zip(snake_case_ , snake_case_ ): self.assertAlmostEqual(snake_case_ , snake_case_ , delta=snake_case_ , msg=snake_case_ ) def _A( self ): lowercase ={'''num_warmup_steps''': 2, '''num_training_steps''': 10} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) lowercase ={ get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {'''num_warmup_steps''': 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, '''num_cycles''': 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, '''power''': 2.0, '''lr_end''': 1E-7}, [0.0, 5.0, 10.0, 7.6_56, 5.6_25, 3.9_06, 2.5, 1.4_06, 0.6_25, 0.1_56], ), get_inverse_sqrt_schedule: ( {'''num_warmup_steps''': 2}, [0.0, 5.0, 10.0, 8.1_65, 7.0_71, 6.3_25, 5.7_74, 5.3_45, 5.0, 4.7_14], ), } for scheduler_func, data in scheds.items(): lowercase , lowercase =data lowercase =scheduler_func(self.optimizer , **snake_case_ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) lowercase =unwrap_schedule(snake_case_ , self.num_steps ) self.assertListAlmostEqual( snake_case_ , snake_case_ , tol=1E-2 , msg=f'failed for {scheduler_func} in normal scheduler' , ) lowercase =scheduler_func(self.optimizer , **snake_case_ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(snake_case_ ) # wrap to test picklability of the schedule lowercase =unwrap_and_save_reload_schedule(snake_case_ , self.num_steps ) self.assertListEqual(snake_case_ , snake_case_ , msg=f'failed for {scheduler_func} in save and reload' ) class __magic_name__ : def __init__( self , snake_case_ ): lowercase =fn def __call__( self , *snake_case_ , **snake_case_ ): return self.fn(*snake_case_ , **snake_case_ ) @classmethod def _A( self , snake_case_ ): lowercase =list(map(self , scheduler.lr_lambdas ) )
72
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = ["pixel_values"] def __init__( self : Union[str, Any] , A__ : bool = True , A__ : Dict[str, int] = None , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = True , **A__ : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : str = size if size is not None else {"shortest_edge": 2_24} snake_case_ : Union[str, Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : str = size snake_case_ : Optional[Any] = resample snake_case_ : Any = do_center_crop snake_case_ : Any = crop_size snake_case_ : str = do_rescale snake_case_ : Optional[Any] = rescale_factor snake_case_ : int = do_normalize snake_case_ : Optional[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ : List[str] = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ : int = do_convert_rgb def UpperCAmelCase__ ( self : Optional[int] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : str = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : np.ndarray , A__ : Union[int, float] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> str: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Any , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Any , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[Any] , A__ : ImageInput , A__ : bool = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : int = None , A__ : bool = None , A__ : float = None , A__ : bool = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A__ : Optional[Any] , ) -> PIL.Image.Image: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Union[str, Any] = size if size is not None else self.size snake_case_ : Any = get_size_dict(A__ , param_name="size" , default_to_square=A__ ) snake_case_ : Optional[int] = resample if resample is not None else self.resample snake_case_ : int = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" , default_to_square=A__ ) snake_case_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : List[Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ : List[Any] = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ : Dict = [convert_to_rgb(A__ ) for image in images] # All transformations expect numpy arrays. snake_case_ : Dict = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : Tuple = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : int = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : List[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Tuple = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ )
666
0
a_ : str = 'Alexander Joslin' import operator as op from .stack import Stack def lowerCamelCase__ (_UpperCAmelCase): SCREAMING_SNAKE_CASE = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} SCREAMING_SNAKE_CASE = Stack() SCREAMING_SNAKE_CASE = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(_UpperCAmelCase)) elif i in operators: # RULE 2 operator_stack.push(_UpperCAmelCase) elif i == ")": # RULE 4 SCREAMING_SNAKE_CASE = operator_stack.peek() operator_stack.pop() SCREAMING_SNAKE_CASE = operand_stack.peek() operand_stack.pop() SCREAMING_SNAKE_CASE = operand_stack.peek() operand_stack.pop() SCREAMING_SNAKE_CASE = operators[opr](_UpperCAmelCase , _UpperCAmelCase) operand_stack.push(_UpperCAmelCase) # RULE 5 return operand_stack.peek() if __name__ == "__main__": a_ : Dict = '(5 + ((4 * 2) * (2 + 3)))' # answer = 45 print(f"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
73
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): snake_case_ ,snake_case_ : Dict = position snake_case_ : int = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] snake_case_ : Union[str, Any] = [] for position in positions: snake_case_ ,snake_case_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(lowerCAmelCase_ ) return permissible_positions def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] , lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): if is_complete(lowerCAmelCase_ ): return True for position in get_valid_pos(lowerCAmelCase_ , len(lowerCAmelCase_ ) ): snake_case_ ,snake_case_ : Dict = position if board[y][x] == 0: snake_case_ : List[str] = curr + 1 if open_knight_tour_helper(lowerCAmelCase_ , lowerCAmelCase_ , curr + 1 ): return True snake_case_ : Dict = 0 return False def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): snake_case_ : Optional[Any] = 1 if open_knight_tour_helper(lowerCAmelCase_ , (i, j) , 1 ): return board snake_case_ : Dict = 0 snake_case_ : str = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
666
0
def a__ ( snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = [] if len(snake_case ) == 1: return [nums.copy()] for _ in range(len(snake_case ) ): __SCREAMING_SNAKE_CASE : Optional[int] = nums.pop(0 ) __SCREAMING_SNAKE_CASE : int = permute(snake_case ) for perm in permutations: perm.append(snake_case ) result.extend(snake_case ) nums.append(snake_case ) return result def a__ ( snake_case ): """simple docstring""" def backtrack(snake_case ): if start == len(snake_case ) - 1: output.append(nums[:] ) else: for i in range(snake_case , len(snake_case ) ): __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Dict = nums[i], nums[start] backtrack(start + 1 ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Union[str, Any] = nums[i], nums[start] # backtrack __SCREAMING_SNAKE_CASE : Optional[Any] = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function lowercase_ = permutea([1, 2, 3]) print(res) doctest.testmod()
74
from ...configuration_utils import PretrainedConfig class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = "bert-generation" def __init__( self : Optional[int] , A__ : List[Any]=5_03_58 , A__ : Any=10_24 , A__ : Any=24 , A__ : List[Any]=16 , A__ : List[Any]=40_96 , A__ : int="gelu" , A__ : List[str]=0.1 , A__ : List[str]=0.1 , A__ : str=5_12 , A__ : int=0.02 , A__ : Any=1E-12 , A__ : Optional[Any]=0 , A__ : List[str]=2 , A__ : Optional[int]=1 , A__ : str="absolute" , A__ : Any=True , **A__ : Optional[Any] , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , **A__ ) snake_case_ : str = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Optional[Any] = hidden_act snake_case_ : Tuple = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : Optional[Any] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = position_embedding_type snake_case_ : Dict = use_cache
666
0
'''simple docstring''' import argparse import os import re UpperCamelCase__ = '''src/diffusers''' # Pattern that looks at the indentation in a line. UpperCamelCase__ = re.compile(R'''^(\s*)\S''') # Pattern that matches `"key":" and puts `key` in group 0. UpperCamelCase__ = re.compile(R'''^\s*"([^"]+)":''') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. UpperCamelCase__ = re.compile(R'''^\s*_import_structure\["([^"]+)"\]''') # Pattern that matches `"key",` and puts `key` in group 0. UpperCamelCase__ = re.compile(R'''^\s*"([^"]+)",\s*$''') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. UpperCamelCase__ = re.compile(R'''\[([^\]]+)\]''') def a__ ( lowerCAmelCase__ ) -> Optional[Any]: UpperCAmelCase__ : Optional[int] = _re_indent.search(lowerCAmelCase__ ) return "" if search is None else search.groups()[0] def a__ ( lowerCAmelCase__ , lowerCAmelCase__="" , lowerCAmelCase__=None , lowerCAmelCase__=None ) -> List[Any]: UpperCAmelCase__ : int = 0 UpperCAmelCase__ : Union[str, Any] = code.split('''\n''' ) if start_prompt is not None: while not lines[index].startswith(lowerCAmelCase__ ): index += 1 UpperCAmelCase__ : Any = ['''\n'''.join(lines[:index] )] else: UpperCAmelCase__ : Union[str, Any] = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). UpperCAmelCase__ : int = [lines[index]] index += 1 while index < len(lowerCAmelCase__ ) and (end_prompt is None or not lines[index].startswith(lowerCAmelCase__ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(lowerCAmelCase__ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ''' ''' ): current_block.append(lines[index] ) blocks.append('''\n'''.join(lowerCAmelCase__ ) ) if index < len(lowerCAmelCase__ ) - 1: UpperCAmelCase__ : str = [lines[index + 1]] index += 1 else: UpperCAmelCase__ : Optional[int] = [] else: blocks.append('''\n'''.join(lowerCAmelCase__ ) ) UpperCAmelCase__ : Dict = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(lowerCAmelCase__ ) > 0: blocks.append('''\n'''.join(lowerCAmelCase__ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(lowerCAmelCase__ ): blocks.append('''\n'''.join(lines[index:] ) ) return blocks def a__ ( lowerCAmelCase__ ) -> List[Any]: def _inner(lowerCAmelCase__ ): return key(lowerCAmelCase__ ).lower().replace('''_''' , '''''' ) return _inner def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ) -> int: # If no key is provided, we use a noop. def noop(lowerCAmelCase__ ): return x if key is None: UpperCAmelCase__ : Tuple = noop # Constants are all uppercase, they go first. UpperCAmelCase__ : Optional[Any] = [obj for obj in objects if key(lowerCAmelCase__ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. UpperCAmelCase__ : int = [obj for obj in objects if key(lowerCAmelCase__ )[0].isupper() and not key(lowerCAmelCase__ ).isupper()] # Functions begin with a lowercase, they go last. UpperCAmelCase__ : Dict = [obj for obj in objects if not key(lowerCAmelCase__ )[0].isupper()] UpperCAmelCase__ : Optional[Any] = ignore_underscore(lowerCAmelCase__ ) return sorted(lowerCAmelCase__ , key=lowerCAmelCase__ ) + sorted(lowerCAmelCase__ , key=lowerCAmelCase__ ) + sorted(lowerCAmelCase__ , key=lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ ) -> Any: # This inner function sort imports between [ ]. def _replace(lowerCAmelCase__ ): UpperCAmelCase__ : Optional[int] = match.groups()[0] if "," not in imports: return F"""[{imports}]""" UpperCAmelCase__ : Optional[int] = [part.strip().replace('''"''' , '''''' ) for part in imports.split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase__ : Tuple = keys[:-1] return "[" + ", ".join([F"""\"{k}\"""" for k in sort_objects(lowerCAmelCase__ )] ) + "]" UpperCAmelCase__ : Tuple = import_statement.split('''\n''' ) if len(lowerCAmelCase__ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. UpperCAmelCase__ : Dict = 2 if lines[1].strip() == '''[''' else 1 UpperCAmelCase__ : int = [(i, _re_strip_line.search(lowerCAmelCase__ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] UpperCAmelCase__ : List[str] = sort_objects(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : x[1] ) UpperCAmelCase__ : Optional[Any] = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(lowerCAmelCase__ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: UpperCAmelCase__ : Optional[Any] = _re_bracket_content.sub(_replace , lines[1] ) else: UpperCAmelCase__ : Union[str, Any] = [part.strip().replace('''"''' , '''''' ) for part in lines[1].split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCAmelCase__ : List[str] = keys[:-1] UpperCAmelCase__ : str = get_indent(lines[1] ) + ''', '''.join([F"""\"{k}\"""" for k in sort_objects(lowerCAmelCase__ )] ) return "\n".join(lowerCAmelCase__ ) else: # Finally we have to deal with imports fitting on one line UpperCAmelCase__ : Any = _re_bracket_content.sub(_replace , lowerCAmelCase__ ) return import_statement def a__ ( lowerCAmelCase__ , lowerCAmelCase__=True ) -> Tuple: with open(lowerCAmelCase__ , '''r''' ) as f: UpperCAmelCase__ : Dict = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 UpperCAmelCase__ : Tuple = split_code_in_indented_blocks( lowerCAmelCase__ , start_prompt='''_import_structure = {''' , end_prompt='''if TYPE_CHECKING:''' ) # We ignore block 0 (everything until start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(lowerCAmelCase__ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. UpperCAmelCase__ : List[str] = main_blocks[block_idx] UpperCAmelCase__ : Tuple = block.split('''\n''' ) # Get to the start of the imports. UpperCAmelCase__ : List[str] = 0 while line_idx < len(lowerCAmelCase__ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: UpperCAmelCase__ : Dict = len(lowerCAmelCase__ ) else: line_idx += 1 if line_idx >= len(lowerCAmelCase__ ): continue # Ignore beginning and last line: they don't contain anything. UpperCAmelCase__ : List[Any] = '''\n'''.join(block_lines[line_idx:-1] ) UpperCAmelCase__ : Tuple = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. UpperCAmelCase__ : List[Any] = split_code_in_indented_blocks(lowerCAmelCase__ , indent_level=lowerCAmelCase__ ) # We have two categories of import key: list or _import_structure[key].append/extend UpperCAmelCase__ : Tuple = _re_direct_key if '''_import_structure''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. UpperCAmelCase__ : Tuple = [(pattern.search(lowerCAmelCase__ ).groups()[0] if pattern.search(lowerCAmelCase__ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. UpperCAmelCase__ : Any = [(i, key) for i, key in enumerate(lowerCAmelCase__ ) if key is not None] UpperCAmelCase__ : Dict = [x[0] for x in sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. UpperCAmelCase__ : Tuple = 0 UpperCAmelCase__ : Tuple = [] for i in range(len(lowerCAmelCase__ ) ): if keys[i] is None: reordered_blocks.append(internal_blocks[i] ) else: UpperCAmelCase__ : Any = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reordered_blocks.append(lowerCAmelCase__ ) count += 1 # And we put our main block back together with its first and last line. UpperCAmelCase__ : Union[str, Any] = '''\n'''.join(block_lines[:line_idx] + reordered_blocks + [block_lines[-1]] ) if code != "\n".join(lowerCAmelCase__ ): if check_only: return True else: print(F"""Overwriting {file}.""" ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write('''\n'''.join(lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__=True ) -> Optional[Any]: UpperCAmelCase__ : str = [] for root, _, files in os.walk(lowerCAmelCase__ ): if "__init__.py" in files: UpperCAmelCase__ : Optional[int] = sort_imports(os.path.join(lowerCAmelCase__ , '''__init__.py''' ) , check_only=lowerCAmelCase__ ) if result: UpperCAmelCase__ : int = [os.path.join(lowerCAmelCase__ , '''__init__.py''' )] if len(lowerCAmelCase__ ) > 0: raise ValueError(F"""Would overwrite {len(lowerCAmelCase__ )} files, run `make style`.""" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') UpperCamelCase__ = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
75
import math def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [] snake_case_ : List[str] = 2 snake_case_ : Optional[int] = int(math.sqrt(lowerCAmelCase_ ) ) # Size of every segment snake_case_ : str = [True] * (end + 1) snake_case_ : Any = [] while start <= end: if temp[start] is True: in_prime.append(lowerCAmelCase_ ) for i in range(start * start , end + 1 , lowerCAmelCase_ ): snake_case_ : Union[str, Any] = False start += 1 prime += in_prime snake_case_ : Dict = end + 1 snake_case_ : Dict = min(2 * end , lowerCAmelCase_ ) while low <= n: snake_case_ : Any = [True] * (high - low + 1) for each in in_prime: snake_case_ : Optional[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(lowerCAmelCase_ , high + 1 , lowerCAmelCase_ ): snake_case_ : List[Any] = False for j in range(len(lowerCAmelCase_ ) ): if temp[j] is True: prime.append(j + low ) snake_case_ : int = high + 1 snake_case_ : Union[str, Any] = min(high + end , lowerCAmelCase_ ) return prime print(sieve(1_0**6))
666
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="open-llama" def __init__( self , UpperCamelCase_=10_00_00 , UpperCamelCase_=40_96 , UpperCamelCase_=1_10_08 , UpperCamelCase_=32 , UpperCamelCase_=32 , UpperCamelCase_="silu" , UpperCamelCase_=20_48 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1E-6 , UpperCamelCase_=True , UpperCamelCase_=0 , UpperCamelCase_=1 , UpperCamelCase_=2 , UpperCamelCase_=False , UpperCamelCase_=True , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=True , UpperCamelCase_=True , UpperCamelCase_=None , **UpperCamelCase_ , ) -> Union[str, Any]: __lowercase : Any = vocab_size __lowercase : Tuple = max_position_embeddings __lowercase : Optional[int] = hidden_size __lowercase : Any = intermediate_size __lowercase : List[Any] = num_hidden_layers __lowercase : str = num_attention_heads __lowercase : Optional[Any] = hidden_act __lowercase : Optional[int] = initializer_range __lowercase : Optional[Any] = rms_norm_eps __lowercase : Tuple = use_cache __lowercase : Optional[Any] = kwargs.pop( '''use_memorry_efficient_attention''' , UpperCamelCase_ ) __lowercase : Optional[Any] = hidden_dropout_prob __lowercase : Optional[int] = attention_dropout_prob __lowercase : List[Any] = use_stable_embedding __lowercase : Union[str, Any] = shared_input_output_embedding __lowercase : int = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , tie_word_embeddings=UpperCamelCase_ , **UpperCamelCase_ , ) def _lowerCamelCase ( self ) -> Tuple: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , UpperCamelCase_ ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""" ) __lowercase : Dict = self.rope_scaling.get('''type''' , UpperCamelCase_ ) __lowercase : Any = self.rope_scaling.get('''factor''' , UpperCamelCase_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(UpperCamelCase_ , UpperCamelCase_ ) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
76
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self : List[str] , A__ : List[Any] , A__ : int=7 , A__ : Union[str, Any]=3 , A__ : List[str]=30 , A__ : Optional[int]=4_00 , A__ : Optional[Any]=True , A__ : Optional[int]=None , A__ : Optional[Any]=True , A__ : Any=[0.5, 0.5, 0.5] , A__ : int=[0.5, 0.5, 0.5] , A__ : Any=True , A__ : int=1 / 2_55 , A__ : List[str]=True , ) -> Dict: '''simple docstring''' snake_case_ : int = size if size is not None else {"shortest_edge": 18, "longest_edge": 13_33} snake_case_ : Any = parent snake_case_ : Optional[int] = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Union[str, Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Tuple = do_resize snake_case_ : Dict = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : List[Any] = image_std snake_case_ : Tuple = do_rescale snake_case_ : Any = rescale_factor snake_case_ : Optional[int] = do_pad def UpperCAmelCase__ ( self : int ) -> List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase__ ( self : Optional[int] , A__ : Optional[int] , A__ : Any=False ) -> Optional[Any]: '''simple docstring''' if not batched: snake_case_ : Any = image_inputs[0] if isinstance(A__ , Image.Image ): snake_case_ ,snake_case_ : Dict = image.size else: snake_case_ ,snake_case_ : int = image.shape[1], image.shape[2] if w < h: snake_case_ : Dict = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[int] = self.size["shortest_edge"] elif w > h: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : str = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : List[Any] = self.size["shortest_edge"] else: snake_case_ : str = [] for image in image_inputs: snake_case_ ,snake_case_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : List[Any] = max(A__ , key=lambda A__ : item[0] )[0] snake_case_ : int = max(A__ , key=lambda A__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( _UpperCamelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[int] = ConditionalDetrImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' snake_case_ : List[str] = ConditionalDetrImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Any ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A__ , "image_mean" ) ) self.assertTrue(hasattr(A__ , "image_std" ) ) self.assertTrue(hasattr(A__ , "do_normalize" ) ) self.assertTrue(hasattr(A__ , "do_resize" ) ) self.assertTrue(hasattr(A__ , "size" ) ) def UpperCAmelCase__ ( self : List[str] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 13_33} ) self.assertEqual(image_processor.do_pad , A__ ) snake_case_ : Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A__ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , A__ ) def UpperCAmelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' pass def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ ) for image in image_inputs: self.assertIsInstance(A__ , Image.Image ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) snake_case_ : int = image_processing(A__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : int ) -> Any: '''simple docstring''' snake_case_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , numpify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , np.ndarray ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[str] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Optional[int] = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Dict = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : Tuple ) -> str: '''simple docstring''' snake_case_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , torchify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , torch.Tensor ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Any = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : int = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def UpperCAmelCase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : Optional[Any] = json.loads(f.read() ) snake_case_ : int = {"image_id": 3_97_69, "annotations": target} # encode them snake_case_ : Optional[int] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) snake_case_ : Any = image_processing(images=A__ , annotations=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : List[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : List[str] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Tuple = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify orig_size snake_case_ : Any = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) ) @slow def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' snake_case_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : Any = json.loads(f.read() ) snake_case_ : Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_97_69, "segments_info": target} snake_case_ : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor(format="coco_panoptic" ) snake_case_ : str = image_processing(images=A__ , annotations=A__ , masks_path=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : int = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Optional[int] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Optional[int] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify masks snake_case_ : Union[str, Any] = 82_28_73 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , A__ ) # verify orig_size snake_case_ : Dict = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : str = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) )
666
0
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class a__ ( __magic_name__ ): lowercase_ = 42 class a__ ( __magic_name__ , __magic_name__ ): @register_to_config def __init__( self : Optional[int] , UpperCamelCase_ : int = 32 , UpperCamelCase_ : int = 64 , UpperCamelCase_ : int = 20 , UpperCamelCase_ : int = 768 , UpperCamelCase_ : List[Any]=77 , UpperCamelCase_ : Dict=4 , UpperCamelCase_ : float = 0.0 , UpperCamelCase_ : str = "silu" , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = "linear" , UpperCamelCase_ : Optional[str] = "prd" , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , ): """simple docstring""" super().__init__() __UpperCAmelCase : int = num_attention_heads __UpperCAmelCase : int = attention_head_dim __UpperCAmelCase : Tuple = num_attention_heads * attention_head_dim __UpperCAmelCase : List[str] = additional_embeddings __UpperCAmelCase : int = time_embed_dim or inner_dim __UpperCAmelCase : List[str] = embedding_proj_dim or embedding_dim __UpperCAmelCase : Optional[Any] = clip_embed_dim or embedding_dim __UpperCAmelCase : Tuple = Timesteps(UpperCamelCase_ , UpperCamelCase_ , 0) __UpperCAmelCase : List[str] = TimestepEmbedding(UpperCamelCase_ , UpperCamelCase_ , out_dim=UpperCamelCase_ , act_fn=UpperCamelCase_) __UpperCAmelCase : int = nn.Linear(UpperCamelCase_ , UpperCamelCase_) if embedding_proj_norm_type is None: __UpperCAmelCase : Union[str, Any] = None elif embedding_proj_norm_type == "layer": __UpperCAmelCase : str = nn.LayerNorm(UpperCamelCase_) else: raise ValueError(F"unsupported embedding_proj_norm_type: {embedding_proj_norm_type}") __UpperCAmelCase : List[str] = nn.Linear(UpperCamelCase_ , UpperCamelCase_) if encoder_hid_proj_type is None: __UpperCAmelCase : List[str] = None elif encoder_hid_proj_type == "linear": __UpperCAmelCase : List[str] = nn.Linear(UpperCamelCase_ , UpperCamelCase_) else: raise ValueError(F"unsupported encoder_hid_proj_type: {encoder_hid_proj_type}") __UpperCAmelCase : List[Any] = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , UpperCamelCase_)) if added_emb_type == "prd": __UpperCAmelCase : Union[str, Any] = nn.Parameter(torch.zeros(1 , 1 , UpperCamelCase_)) elif added_emb_type is None: __UpperCAmelCase : int = None else: raise ValueError( F"`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`.") __UpperCAmelCase : Union[str, Any] = nn.ModuleList( [ BasicTransformerBlock( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , dropout=UpperCamelCase_ , activation_fn="gelu" , attention_bias=UpperCamelCase_ , ) for d in range(UpperCamelCase_) ]) if norm_in_type == "layer": __UpperCAmelCase : List[Any] = nn.LayerNorm(UpperCamelCase_) elif norm_in_type is None: __UpperCAmelCase : Any = None else: raise ValueError(F"Unsupported norm_in_type: {norm_in_type}.") __UpperCAmelCase : List[str] = nn.LayerNorm(UpperCamelCase_) __UpperCAmelCase : Any = nn.Linear(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Any = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0) causal_attention_mask.triu_(1) __UpperCAmelCase : str = causal_attention_mask[None, ...] self.register_buffer("causal_attention_mask" , UpperCamelCase_ , persistent=UpperCamelCase_) __UpperCAmelCase : Tuple = nn.Parameter(torch.zeros(1 , UpperCamelCase_)) __UpperCAmelCase : Any = nn.Parameter(torch.zeros(1 , UpperCamelCase_)) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Dict = {} def fn_recursive_add_processors(UpperCamelCase_ : str , UpperCamelCase_ : torch.nn.Module , UpperCamelCase_ : Dict[str, AttentionProcessor]): if hasattr(UpperCamelCase_ , "set_processor"): __UpperCAmelCase : int = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"{name}.{sub_name}" , UpperCamelCase_ , UpperCamelCase_) return processors for name, module in self.named_children(): fn_recursive_add_processors(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) return processors def a_ ( self : str , UpperCamelCase_ : Union[AttentionProcessor, Dict[str, AttentionProcessor]]): """simple docstring""" __UpperCAmelCase : int = len(self.attn_processors.keys()) if isinstance(UpperCamelCase_ , UpperCamelCase_) and len(UpperCamelCase_) != count: raise ValueError( F"A dict of processors was passed, but the number of processors {len(UpperCamelCase_)} does not match the" F" number of attention layers: {count}. Please make sure to pass {count} processor classes.") def fn_recursive_attn_processor(UpperCamelCase_ : str , UpperCamelCase_ : torch.nn.Module , UpperCamelCase_ : Any): if hasattr(UpperCamelCase_ , "set_processor"): if not isinstance(UpperCamelCase_ , UpperCamelCase_): module.set_processor(UpperCamelCase_) else: module.set_processor(processor.pop(F"{name}.processor")) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"{name}.{sub_name}" , UpperCamelCase_ , UpperCamelCase_) for name, module in self.named_children(): fn_recursive_attn_processor(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) def a_ ( self : str): """simple docstring""" self.set_attn_processor(AttnProcessor()) def a_ ( self : Dict , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Union[torch.Tensor, float, int] , UpperCamelCase_ : torch.FloatTensor , UpperCamelCase_ : Optional[torch.FloatTensor] = None , UpperCamelCase_ : Optional[torch.BoolTensor] = None , UpperCamelCase_ : bool = True , ): """simple docstring""" __UpperCAmelCase : int = hidden_states.shape[0] __UpperCAmelCase : Union[str, Any] = timestep if not torch.is_tensor(UpperCamelCase_): __UpperCAmelCase : Dict = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device) elif torch.is_tensor(UpperCamelCase_) and len(timesteps.shape) == 0: __UpperCAmelCase : Union[str, Any] = timesteps[None].to(hidden_states.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCAmelCase : str = timesteps * torch.ones(UpperCamelCase_ , dtype=timesteps.dtype , device=timesteps.device) __UpperCAmelCase : Union[str, Any] = self.time_proj(UpperCamelCase_) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. __UpperCAmelCase : Any = timesteps_projected.to(dtype=self.dtype) __UpperCAmelCase : Optional[int] = self.time_embedding(UpperCamelCase_) if self.embedding_proj_norm is not None: __UpperCAmelCase : List[str] = self.embedding_proj_norm(UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = self.embedding_proj(UpperCamelCase_) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: __UpperCAmelCase : Union[str, Any] = self.encoder_hidden_states_proj(UpperCamelCase_) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set") __UpperCAmelCase : List[Any] = self.proj_in(UpperCamelCase_) __UpperCAmelCase : str = self.positional_embedding.to(hidden_states.dtype) __UpperCAmelCase : int = [] __UpperCAmelCase : Union[str, Any] = 0 if encoder_hidden_states is not None: additional_embeds.append(UpperCamelCase_) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape) == 2: __UpperCAmelCase : Union[str, Any] = proj_embeddings[:, None, :] if len(hidden_states.shape) == 2: __UpperCAmelCase : List[str] = hidden_states[:, None, :] __UpperCAmelCase : List[str] = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: __UpperCAmelCase : Union[str, Any] = self.prd_embedding.to(hidden_states.dtype).expand(UpperCamelCase_ , -1 , -1) additional_embeds.append(UpperCamelCase_) __UpperCAmelCase : Optional[int] = torch.cat( UpperCamelCase_ , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens __UpperCAmelCase : List[Any] = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: __UpperCAmelCase : str = F.pad( UpperCamelCase_ , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) __UpperCAmelCase : Tuple = hidden_states + positional_embeddings if attention_mask is not None: __UpperCAmelCase : Dict = (1 - attention_mask.to(hidden_states.dtype)) * -10000.0 __UpperCAmelCase : str = F.pad(UpperCamelCase_ , (0, self.additional_embeddings) , value=0.0) __UpperCAmelCase : List[str] = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype) __UpperCAmelCase : List[Any] = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0) if self.norm_in is not None: __UpperCAmelCase : int = self.norm_in(UpperCamelCase_) for block in self.transformer_blocks: __UpperCAmelCase : int = block(UpperCamelCase_ , attention_mask=UpperCamelCase_) __UpperCAmelCase : Any = self.norm_out(UpperCamelCase_) if self.prd_embedding is not None: __UpperCAmelCase : Dict = hidden_states[:, -1] else: __UpperCAmelCase : str = hidden_states[:, additional_embeddings_len:] __UpperCAmelCase : List[str] = self.proj_to_clip_embeddings(UpperCamelCase_) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=UpperCamelCase_) def a_ ( self : str , UpperCamelCase_ : Dict): """simple docstring""" __UpperCAmelCase : str = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
77
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "Model type selected in the list: " + ", ".join(_UpperCamelCase )} ) _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) _SCREAMING_SNAKE_CASE : int = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) _SCREAMING_SNAKE_CASE : float = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) _SCREAMING_SNAKE_CASE : int = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = "train" _SCREAMING_SNAKE_CASE : Any = "dev" class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : SquadDataTrainingArguments _SCREAMING_SNAKE_CASE : List[SquadFeatures] _SCREAMING_SNAKE_CASE : Split _SCREAMING_SNAKE_CASE : bool def __init__( self : str , A__ : SquadDataTrainingArguments , A__ : PreTrainedTokenizer , A__ : Optional[int] = None , A__ : Union[str, Split] = Split.train , A__ : Optional[bool] = False , A__ : Optional[str] = None , A__ : Optional[str] = "pt" , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = args snake_case_ : int = is_language_sensitive snake_case_ : int = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(A__ , A__ ): try: snake_case_ : List[str] = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) snake_case_ : Tuple = mode # Load data features from cache or dataset file snake_case_ : Dict = "v2" if args.version_2_with_negative else "v1" snake_case_ : List[Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case_ : List[Any] = cached_features_file + ".lock" with FileLock(A__ ): if os.path.exists(A__ ) and not args.overwrite_cache: snake_case_ : int = time.time() snake_case_ : List[Any] = torch.load(A__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. snake_case_ : Tuple = self.old_features["features"] snake_case_ : List[str] = self.old_features.get("dataset" , A__ ) snake_case_ : Tuple = self.old_features.get("examples" , A__ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" " future run" ) else: if mode == Split.dev: snake_case_ : Tuple = self.processor.get_dev_examples(args.data_dir ) else: snake_case_ : Tuple = self.processor.get_train_examples(args.data_dir ) snake_case_ ,snake_case_ : Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=A__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=A__ , ) snake_case_ : Any = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , A__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self : str ) -> Dict: '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[int] , A__ : Optional[int] ) -> Dict[str, torch.Tensor]: '''simple docstring''' snake_case_ : Any = self.features[i] snake_case_ : Optional[int] = torch.tensor(feature.input_ids , dtype=torch.long ) snake_case_ : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.cls_index , dtype=torch.long ) snake_case_ : str = torch.tensor(feature.p_mask , dtype=torch.float ) snake_case_ : str = torch.tensor(feature.is_impossible , dtype=torch.float ) snake_case_ : Optional[int] = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: snake_case_ : Any = torch.tensor(feature.start_position , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
666
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "git_vision_model" def __init__( self : int , A__ : Union[str, Any]=7_68 , A__ : List[Any]=30_72 , A__ : Tuple=12 , A__ : Optional[Any]=12 , A__ : Optional[int]=3 , A__ : List[str]=2_24 , A__ : Dict=16 , A__ : int="quick_gelu" , A__ : Any=1E-5 , A__ : Tuple=0.0 , A__ : Optional[int]=0.02 , **A__ : List[str] , ) -> Optional[int]: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[Any] = hidden_size snake_case_ : str = intermediate_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = num_channels snake_case_ : Union[str, Any] = patch_size snake_case_ : List[str] = image_size snake_case_ : List[Any] = initializer_range snake_case_ : Any = attention_dropout snake_case_ : Any = layer_norm_eps snake_case_ : int = hidden_act @classmethod def UpperCAmelCase__ ( cls : List[Any] , A__ : Union[str, os.PathLike] , **A__ : Optional[int] ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(A__ ) snake_case_ ,snake_case_ : Tuple = cls.get_config_dict(A__ , **A__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": snake_case_ : Any = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(A__ , **A__ ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = "git" def __init__( self : Any , A__ : List[str]=None , A__ : List[str]=3_05_22 , A__ : Tuple=7_68 , A__ : Tuple=6 , A__ : str=12 , A__ : Any=30_72 , A__ : List[str]="gelu" , A__ : int=0.1 , A__ : Dict=0.1 , A__ : Any=10_24 , A__ : Optional[Any]=0.02 , A__ : Optional[Any]=1E-12 , A__ : Dict=0 , A__ : Any="absolute" , A__ : Tuple=True , A__ : Any=False , A__ : Tuple=1_01 , A__ : Tuple=1_02 , A__ : List[Any]=None , **A__ : List[str] , ) -> int: '''simple docstring''' super().__init__(bos_token_id=A__ , eos_token_id=A__ , pad_token_id=A__ , **A__ ) if vision_config is None: snake_case_ : int = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) snake_case_ : str = GitVisionConfig(**A__ ) snake_case_ : int = vocab_size snake_case_ : List[Any] = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : List[Any] = num_attention_heads snake_case_ : Any = hidden_act snake_case_ : Dict = intermediate_size snake_case_ : Any = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = initializer_range snake_case_ : List[str] = layer_norm_eps snake_case_ : Any = position_embedding_type snake_case_ : Union[str, Any] = use_cache snake_case_ : str = tie_word_embeddings snake_case_ : List[Any] = num_image_with_embedding snake_case_ : Dict = bos_token_id snake_case_ : int = eos_token_id def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' snake_case_ : Tuple = copy.deepcopy(self.__dict__ ) snake_case_ : Optional[int] = self.vision_config.to_dict() snake_case_ : Tuple = self.__class__.model_type return output
666
0
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=99 , _lowerCAmelCase=32 , _lowerCAmelCase=5 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=16 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=4 , _lowerCAmelCase=None , ): UpperCAmelCase__ : int = parent UpperCAmelCase__ : Tuple = batch_size UpperCAmelCase__ : Optional[Any] = seq_length UpperCAmelCase__ : str = is_training UpperCAmelCase__ : List[Any] = use_input_mask UpperCAmelCase__ : Tuple = use_token_type_ids UpperCAmelCase__ : Any = use_labels UpperCAmelCase__ : List[str] = vocab_size UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : Optional[int] = num_hidden_layers UpperCAmelCase__ : Union[str, Any] = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : Tuple = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = max_position_embeddings UpperCAmelCase__ : List[str] = type_vocab_size UpperCAmelCase__ : str = type_sequence_label_size UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Optional[int] = num_labels UpperCAmelCase__ : Tuple = num_choices UpperCAmelCase__ : Optional[Any] = scope def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : int = None if self.use_input_mask: UpperCAmelCase__ : int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Optional[int] = None if self.use_token_type_ids: UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Optional[Any] = None UpperCAmelCase__ : Tuple = None if self.use_labels: UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ : str = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Any = BioGptModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase__ : Dict = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ): UpperCAmelCase__ : Dict = BioGptForCausalLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase__ : Optional[Any] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , *_lowerCAmelCase ): UpperCAmelCase__ : Tuple = BioGptModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() # create attention mask UpperCAmelCase__ : Dict = torch.ones(input_ids.shape , dtype=torch.long , device=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = self.seq_length // 2 UpperCAmelCase__ : List[Any] = 0 # first forward pass UpperCAmelCase__ , UpperCAmelCase__ : Tuple = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase ).to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase__ : Union[str, Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids UpperCAmelCase__ : Union[str, Any] = ids_tensor((1,) , _lowerCAmelCase ).item() + 1 UpperCAmelCase__ : List[Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) UpperCAmelCase__ : str = random_other_next_tokens # append to next input_ids and attn_mask UpperCAmelCase__ : Dict = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase__ : Tuple = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=_lowerCAmelCase )] , dim=1 , ) # get two different outputs UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )["""last_hidden_state"""] UpperCAmelCase__ : Optional[int] = model(_lowerCAmelCase , past_key_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase )["""last_hidden_state"""] # select random slice UpperCAmelCase__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() UpperCAmelCase__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , *_lowerCAmelCase ): UpperCAmelCase__ : Tuple = BioGptModel(config=_lowerCAmelCase ).to(_lowerCAmelCase ).eval() UpperCAmelCase__ : Optional[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=_lowerCAmelCase ) # first forward pass UpperCAmelCase__ : Optional[int] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase__ : Optional[int] = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and UpperCAmelCase__ : Tuple = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase__ : Tuple = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) UpperCAmelCase__ : Union[str, Any] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )["""last_hidden_state"""] UpperCAmelCase__ : str = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase )[ """last_hidden_state""" ] # select random slice UpperCAmelCase__ : str = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase__ : Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase__ : Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , *_lowerCAmelCase , _lowerCAmelCase=False ): UpperCAmelCase__ : Any = BioGptForCausalLM(_lowerCAmelCase ) model.to(_lowerCAmelCase ) if gradient_checkpointing: model.gradient_checkpointing_enable() UpperCAmelCase__ : Tuple = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def __UpperCAmelCase ( self , _lowerCAmelCase , *_lowerCAmelCase ): UpperCAmelCase__ : Any = BioGptModel(_lowerCAmelCase ) UpperCAmelCase__ : str = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_0_1 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.0_1 ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , *_lowerCAmelCase ): UpperCAmelCase__ : Dict = self.num_labels UpperCAmelCase__ : int = BioGptForTokenClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase__ : Any = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Optional[int] = config_and_inputs UpperCAmelCase__ : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) __lowerCamelCase = (BioGptForCausalLM,) if is_torch_available() else () __lowerCamelCase = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = BioGptModelTester(self ) UpperCAmelCase__ : List[str] = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ : Optional[int] = type self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*_lowerCAmelCase , gradient_checkpointing=_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Union[str, Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase__ : Tuple = """left""" # Define PAD Token = EOS Token = 50256 UpperCAmelCase__ : Dict = tokenizer.eos_token UpperCAmelCase__ : str = model.config.eos_token_id # use different length sentences to test batching UpperCAmelCase__ : Tuple = [ """Hello, my dog is a little""", """Today, I""", ] UpperCAmelCase__ : List[Any] = tokenizer(_lowerCAmelCase , return_tensors="""pt""" , padding=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = inputs["""input_ids"""].to(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = model.generate( input_ids=_lowerCAmelCase , attention_mask=inputs["""attention_mask"""].to(_lowerCAmelCase ) , ) UpperCAmelCase__ : Tuple = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = model.generate(input_ids=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() UpperCAmelCase__ : Optional[int] = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = model.generate(input_ids=_lowerCAmelCase , max_length=model.config.max_length - num_paddings ) UpperCAmelCase__ : Union[str, Any] = tokenizer.batch_decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Dict = tokenizer.decode(output_non_padded[0] , skip_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : int = tokenizer.decode(output_padded[0] , skip_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[str] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , [non_padded_sentence, padded_sentence] ) @slow def __UpperCAmelCase ( self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : Dict = BioGptModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : int = 3 UpperCAmelCase__ : int = input_dict["""input_ids"""] UpperCAmelCase__ : int = input_ids.ne(1 ).to(_lowerCAmelCase ) UpperCAmelCase__ : Dict = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = BioGptForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase__ : Optional[int] = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : List[str] = 3 UpperCAmelCase__ : Tuple = """multi_label_classification""" UpperCAmelCase__ : Optional[int] = input_dict["""input_ids"""] UpperCAmelCase__ : int = input_ids.ne(1 ).to(_lowerCAmelCase ) UpperCAmelCase__ : Dict = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) UpperCAmelCase__ : str = BioGptForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() UpperCAmelCase__ : str = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , labels=_lowerCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class UpperCAmelCase_ ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase__ : List[Any] = torch.tensor([[2, 4805, 9, 656, 21]] ) UpperCAmelCase__ : Optional[int] = model(_lowerCAmelCase )[0] UpperCAmelCase__ : str = 42384 UpperCAmelCase__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = torch.tensor( [[[-9.5_2_3_6, -9.8_9_1_8, 1_0.4_5_5_7], [-1_1.0_4_6_9, -9.6_4_2_3, 8.1_0_2_2], [-8.8_6_6_4, -7.8_8_2_6, 5.5_3_2_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1e-4 ) ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase__ : Optional[int] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(_lowerCAmelCase ) torch.manual_seed(0 ) UpperCAmelCase__ : Union[str, Any] = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(_lowerCAmelCase ) UpperCAmelCase__ : Tuple = model.generate( **_lowerCAmelCase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=_lowerCAmelCase , ) UpperCAmelCase__ : Optional[Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : int = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase )
79
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: str ): def get_matched_characters(lowerCAmelCase_: str , lowerCAmelCase_: str ) -> str: snake_case_ : Tuple = [] snake_case_ : Tuple = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): snake_case_ : str = int(max(0 , i - limit ) ) snake_case_ : Optional[int] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) snake_case_ : List[Any] = f"{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}" return "".join(lowerCAmelCase_ ) # matching characters snake_case_ : List[Any] = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : int = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = len(lowerCAmelCase_ ) # transposition snake_case_ : List[str] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: snake_case_ : str = 0.0 else: snake_case_ : Optional[Any] = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters snake_case_ : Optional[Any] = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("hello", "world"))
666
0
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[str] = OrderedDict( [ # Base model mapping ("""albert""", """FlaxAlbertModel"""), ("""bart""", """FlaxBartModel"""), ("""beit""", """FlaxBeitModel"""), ("""bert""", """FlaxBertModel"""), ("""big_bird""", """FlaxBigBirdModel"""), ("""blenderbot""", """FlaxBlenderbotModel"""), ("""blenderbot-small""", """FlaxBlenderbotSmallModel"""), ("""clip""", """FlaxCLIPModel"""), ("""distilbert""", """FlaxDistilBertModel"""), ("""electra""", """FlaxElectraModel"""), ("""gpt-sw3""", """FlaxGPT2Model"""), ("""gpt2""", """FlaxGPT2Model"""), ("""gpt_neo""", """FlaxGPTNeoModel"""), ("""gptj""", """FlaxGPTJModel"""), ("""longt5""", """FlaxLongT5Model"""), ("""marian""", """FlaxMarianModel"""), ("""mbart""", """FlaxMBartModel"""), ("""mt5""", """FlaxMT5Model"""), ("""opt""", """FlaxOPTModel"""), ("""pegasus""", """FlaxPegasusModel"""), ("""regnet""", """FlaxRegNetModel"""), ("""resnet""", """FlaxResNetModel"""), ("""roberta""", """FlaxRobertaModel"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormModel"""), ("""roformer""", """FlaxRoFormerModel"""), ("""t5""", """FlaxT5Model"""), ("""vision-text-dual-encoder""", """FlaxVisionTextDualEncoderModel"""), ("""vit""", """FlaxViTModel"""), ("""wav2vec2""", """FlaxWav2Vec2Model"""), ("""whisper""", """FlaxWhisperModel"""), ("""xglm""", """FlaxXGLMModel"""), ("""xlm-roberta""", """FlaxXLMRobertaModel"""), ] ) __UpperCamelCase : Tuple = OrderedDict( [ # Model for pre-training mapping ("""albert""", """FlaxAlbertForPreTraining"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForPreTraining"""), ("""big_bird""", """FlaxBigBirdForPreTraining"""), ("""electra""", """FlaxElectraForPreTraining"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ("""wav2vec2""", """FlaxWav2Vec2ForPreTraining"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Masked LM mapping ("""albert""", """FlaxAlbertForMaskedLM"""), ("""bart""", """FlaxBartForConditionalGeneration"""), ("""bert""", """FlaxBertForMaskedLM"""), ("""big_bird""", """FlaxBigBirdForMaskedLM"""), ("""distilbert""", """FlaxDistilBertForMaskedLM"""), ("""electra""", """FlaxElectraForMaskedLM"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""roberta""", """FlaxRobertaForMaskedLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMaskedLM"""), ("""roformer""", """FlaxRoFormerForMaskedLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForMaskedLM"""), ] ) __UpperCamelCase : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("""bart""", """FlaxBartForConditionalGeneration"""), ("""blenderbot""", """FlaxBlenderbotForConditionalGeneration"""), ("""blenderbot-small""", """FlaxBlenderbotSmallForConditionalGeneration"""), ("""encoder-decoder""", """FlaxEncoderDecoderModel"""), ("""longt5""", """FlaxLongT5ForConditionalGeneration"""), ("""marian""", """FlaxMarianMTModel"""), ("""mbart""", """FlaxMBartForConditionalGeneration"""), ("""mt5""", """FlaxMT5ForConditionalGeneration"""), ("""pegasus""", """FlaxPegasusForConditionalGeneration"""), ("""t5""", """FlaxT5ForConditionalGeneration"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Image-classsification ("""beit""", """FlaxBeitForImageClassification"""), ("""regnet""", """FlaxRegNetForImageClassification"""), ("""resnet""", """FlaxResNetForImageClassification"""), ("""vit""", """FlaxViTForImageClassification"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ ("""vision-encoder-decoder""", """FlaxVisionEncoderDecoderModel"""), ] ) __UpperCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("""bart""", """FlaxBartForCausalLM"""), ("""bert""", """FlaxBertForCausalLM"""), ("""big_bird""", """FlaxBigBirdForCausalLM"""), ("""electra""", """FlaxElectraForCausalLM"""), ("""gpt-sw3""", """FlaxGPT2LMHeadModel"""), ("""gpt2""", """FlaxGPT2LMHeadModel"""), ("""gpt_neo""", """FlaxGPTNeoForCausalLM"""), ("""gptj""", """FlaxGPTJForCausalLM"""), ("""opt""", """FlaxOPTForCausalLM"""), ("""roberta""", """FlaxRobertaForCausalLM"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForCausalLM"""), ("""xglm""", """FlaxXGLMForCausalLM"""), ("""xlm-roberta""", """FlaxXLMRobertaForCausalLM"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ("""albert""", """FlaxAlbertForSequenceClassification"""), ("""bart""", """FlaxBartForSequenceClassification"""), ("""bert""", """FlaxBertForSequenceClassification"""), ("""big_bird""", """FlaxBigBirdForSequenceClassification"""), ("""distilbert""", """FlaxDistilBertForSequenceClassification"""), ("""electra""", """FlaxElectraForSequenceClassification"""), ("""mbart""", """FlaxMBartForSequenceClassification"""), ("""roberta""", """FlaxRobertaForSequenceClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForSequenceClassification"""), ("""roformer""", """FlaxRoFormerForSequenceClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForSequenceClassification"""), ] ) __UpperCamelCase : List[str] = OrderedDict( [ # Model for Question Answering mapping ("""albert""", """FlaxAlbertForQuestionAnswering"""), ("""bart""", """FlaxBartForQuestionAnswering"""), ("""bert""", """FlaxBertForQuestionAnswering"""), ("""big_bird""", """FlaxBigBirdForQuestionAnswering"""), ("""distilbert""", """FlaxDistilBertForQuestionAnswering"""), ("""electra""", """FlaxElectraForQuestionAnswering"""), ("""mbart""", """FlaxMBartForQuestionAnswering"""), ("""roberta""", """FlaxRobertaForQuestionAnswering"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForQuestionAnswering"""), ("""roformer""", """FlaxRoFormerForQuestionAnswering"""), ("""xlm-roberta""", """FlaxXLMRobertaForQuestionAnswering"""), ] ) __UpperCamelCase : int = OrderedDict( [ # Model for Token Classification mapping ("""albert""", """FlaxAlbertForTokenClassification"""), ("""bert""", """FlaxBertForTokenClassification"""), ("""big_bird""", """FlaxBigBirdForTokenClassification"""), ("""distilbert""", """FlaxDistilBertForTokenClassification"""), ("""electra""", """FlaxElectraForTokenClassification"""), ("""roberta""", """FlaxRobertaForTokenClassification"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForTokenClassification"""), ("""roformer""", """FlaxRoFormerForTokenClassification"""), ("""xlm-roberta""", """FlaxXLMRobertaForTokenClassification"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ # Model for Multiple Choice mapping ("""albert""", """FlaxAlbertForMultipleChoice"""), ("""bert""", """FlaxBertForMultipleChoice"""), ("""big_bird""", """FlaxBigBirdForMultipleChoice"""), ("""distilbert""", """FlaxDistilBertForMultipleChoice"""), ("""electra""", """FlaxElectraForMultipleChoice"""), ("""roberta""", """FlaxRobertaForMultipleChoice"""), ("""roberta-prelayernorm""", """FlaxRobertaPreLayerNormForMultipleChoice"""), ("""roformer""", """FlaxRoFormerForMultipleChoice"""), ("""xlm-roberta""", """FlaxXLMRobertaForMultipleChoice"""), ] ) __UpperCamelCase : str = OrderedDict( [ ("""bert""", """FlaxBertForNextSentencePrediction"""), ] ) __UpperCamelCase : Optional[int] = OrderedDict( [ ("""speech-encoder-decoder""", """FlaxSpeechEncoderDecoderModel"""), ("""whisper""", """FlaxWhisperForConditionalGeneration"""), ] ) __UpperCamelCase : Dict = OrderedDict( [ ("""whisper""", """FlaxWhisperForAudioClassification"""), ] ) __UpperCamelCase : List[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) __UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) __UpperCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) __UpperCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) __UpperCamelCase : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) __UpperCamelCase : Union[str, Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) __UpperCamelCase : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) __UpperCamelCase : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) __UpperCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) __UpperCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Tuple = FLAX_MODEL_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModel) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Union[str, Any] = FLAX_MODEL_FOR_PRETRAINING_MAPPING __UpperCamelCase : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="""pretraining""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="""causal language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING __UpperCamelCase : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="""masked language modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="""sequence-to-sequence language modeling""", checkpoint_for_example="""t5-base""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="""sequence classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING __UpperCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="""question answering""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :List[Any] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[int] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="""token classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[int] = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING __UpperCamelCase : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="""multiple choice""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :str = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING __UpperCamelCase : int = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="""next sentence prediction""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __UpperCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="""image classification""" ) class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Dict = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING __UpperCamelCase : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="""vision-to-text modeling""") class __UpperCamelCase ( _BaseAutoModelClass ): __snake_case :Optional[Any] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING __UpperCamelCase : str = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="""sequence-to-sequence speech-to-text modeling""" )
80
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) set_seed(7_7_0) UpperCAmelCase = { "c_attn": "att_proj", "c_proj": "out_proj", "c_fc": "in_proj", "transformer.": "", "h.": "layers.", "ln_1": "layernorm_1", "ln_2": "layernorm_2", "ln_f": "layernorm_final", "wpe": "position_embeds_layer", "wte": "input_embeds_layer", } UpperCAmelCase = { "text_small": { "repo_id": "suno/bark", "file_name": "text.pt", }, "coarse_small": { "repo_id": "suno/bark", "file_name": "coarse.pt", }, "fine_small": { "repo_id": "suno/bark", "file_name": "fine.pt", }, "text": { "repo_id": "suno/bark", "file_name": "text_2.pt", }, "coarse": { "repo_id": "suno/bark", "file_name": "coarse_2.pt", }, "fine": { "repo_id": "suno/bark", "file_name": "fine_2.pt", }, } UpperCAmelCase = os.path.dirname(os.path.abspath(__file__)) UpperCAmelCase = os.path.join(os.path.expanduser("~"), ".cache") UpperCAmelCase = os.path.join(os.getenv("XDG_CACHE_HOME", default_cache_dir), "suno", "bark_v0") def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int , lowerCAmelCase_: List[str]=False ): snake_case_ : Union[str, Any] = model_type if use_small: key += "_small" return os.path.join(lowerCAmelCase_ , REMOTE_MODEL_PATHS[key]["file_name"] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: List[str] ): os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) hf_hub_download(repo_id=lowerCAmelCase_ , filename=lowerCAmelCase_ , local_dir=lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Any , lowerCAmelCase_: Dict , lowerCAmelCase_: List[str]=False , lowerCAmelCase_: Dict="text" ): if model_type == "text": snake_case_ : int = BarkSemanticModel snake_case_ : str = BarkSemanticConfig snake_case_ : Optional[Any] = BarkSemanticGenerationConfig elif model_type == "coarse": snake_case_ : str = BarkCoarseModel snake_case_ : Optional[int] = BarkCoarseConfig snake_case_ : Any = BarkCoarseGenerationConfig elif model_type == "fine": snake_case_ : Optional[int] = BarkFineModel snake_case_ : Tuple = BarkFineConfig snake_case_ : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() snake_case_ : Optional[Any] = f"{model_type}_small" if use_small else model_type snake_case_ : Any = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(lowerCAmelCase_ ): logger.info(f"{model_type} model not found, downloading into `{CACHE_DIR}`." ) _download(model_info["repo_id"] , model_info["file_name"] ) snake_case_ : Any = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) # this is a hack snake_case_ : Union[str, Any] = checkpoint["model_args"] if "input_vocab_size" not in model_args: snake_case_ : str = model_args["vocab_size"] snake_case_ : Union[str, Any] = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments snake_case_ : Union[str, Any] = model_args.pop("n_head" ) snake_case_ : int = model_args.pop("n_embd" ) snake_case_ : Any = model_args.pop("n_layer" ) snake_case_ : List[str] = ConfigClass(**checkpoint["model_args"] ) snake_case_ : Optional[Any] = ModelClass(config=lowerCAmelCase_ ) snake_case_ : Tuple = GenerationConfigClass() snake_case_ : List[str] = model_generation_config snake_case_ : Optional[int] = checkpoint["model"] # fixup checkpoint snake_case_ : Optional[int] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(lowerCAmelCase_ ): # replace part of the key with corresponding layer name in HF implementation snake_case_ : Tuple = k[len(lowerCAmelCase_ ) :] for old_layer_name in new_layer_name_dict: snake_case_ : int = new_k.replace(lowerCAmelCase_ , new_layer_name_dict[old_layer_name] ) snake_case_ : int = state_dict.pop(lowerCAmelCase_ ) snake_case_ : Optional[int] = set(state_dict.keys() ) - set(model.state_dict().keys() ) snake_case_ : str = {k for k in extra_keys if not k.endswith(".attn.bias" )} snake_case_ : Any = set(model.state_dict().keys() ) - set(state_dict.keys() ) snake_case_ : List[Any] = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(lowerCAmelCase_ ) != 0: raise ValueError(f"extra keys found: {extra_keys}" ) if len(lowerCAmelCase_ ) != 0: raise ValueError(f"missing keys: {missing_keys}" ) model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) snake_case_ : str = model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) snake_case_ : Union[str, Any] = checkpoint["best_val_loss"].item() logger.info(f"model loaded: {round(n_params/1e6 , 1 )}M params, {round(lowerCAmelCase_ , 3 )} loss" ) model.eval() model.to(lowerCAmelCase_ ) del checkpoint, state_dict return model def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: List[Any] , lowerCAmelCase_: str=False , lowerCAmelCase_: int="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() snake_case_ : int = "cpu" # do conversion on cpu snake_case_ : Optional[Any] = _get_ckpt_path(lowerCAmelCase_ , use_small=lowerCAmelCase_ ) snake_case_ : Tuple = _load_model(lowerCAmelCase_ , lowerCAmelCase_ , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) # load bark initial model snake_case_ : int = _bark_load_model(lowerCAmelCase_ , "cpu" , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) if model_type == "text": snake_case_ : Union[str, Any] = bark_model["model"] if model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model snake_case_ : Optional[Any] = 5 snake_case_ : Optional[int] = 1_0 if model_type in ["text", "coarse"]: snake_case_ : Optional[Any] = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) snake_case_ : str = bark_model(lowerCAmelCase_ )[0] snake_case_ : Tuple = model(lowerCAmelCase_ ) # take last logits snake_case_ : List[str] = output_new_model_total.logits[:, [-1], :] else: snake_case_ : Optional[int] = 3 snake_case_ : str = 8 snake_case_ : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) snake_case_ : Any = model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Union[str, Any] = bark_model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: List[str] , lowerCAmelCase_: Any , lowerCAmelCase_: List[Any] , lowerCAmelCase_: int , lowerCAmelCase_: Optional[Any] , ): snake_case_ : Optional[Any] = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkSemanticConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = BarkCoarseConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[str] = BarkFineConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) snake_case_ : List[str] = BarkSemanticModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkCoarseModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Tuple = BarkFineModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Union[str, Any] = EncodecModel.from_pretrained("facebook/encodec_24khz" ) snake_case_ : Tuple = BarkConfig.from_sub_model_configs( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : List[Any] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) snake_case_ : Optional[int] = BarkModel(lowerCAmelCase_ ) snake_case_ : int = semantic snake_case_ : List[str] = coarseAcoustic snake_case_ : str = fineAcoustic snake_case_ : Optional[Any] = codec snake_case_ : Any = bark_generation_config Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) bark.save_pretrained(lowerCAmelCase_ , repo_id=lowerCAmelCase_ , push_to_hub=lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("model_type", type=str, help="text, coarse or fine.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--is_small", action="store_true", help="convert the small version instead of the large.") UpperCAmelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
666
0
import math def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = len(__lowerCamelCase ) __snake_case : Tuple = int(math.floor(math.sqrt(__lowerCamelCase ) ) ) __snake_case : Optional[Any] = 0 while arr[min(__lowerCamelCase , __lowerCamelCase ) - 1] < x: __snake_case : Optional[int] = step step += int(math.floor(math.sqrt(__lowerCamelCase ) ) ) if prev >= n: return -1 while arr[prev] < x: __snake_case : Optional[int] = prev + 1 if prev == min(__lowerCamelCase , __lowerCamelCase ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": _snake_case : int = input("Enter numbers separated by a comma:\n").strip() _snake_case : Tuple = [int(item) for item in user_input.split(",")] _snake_case : str = int(input("Enter the number to be searched:\n")) _snake_case : List[str] = jump_search(arr, x) if res == -1: print("Number not found!") else: print(f'''Number {x} is at index {res}''')
81
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
"""simple docstring""" import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def lowercase__ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_UpperCAmelCase , "tf_padding" ) ) self.parent.assertTrue(hasattr(_UpperCAmelCase , "depth_multiplier" ) ) class lowercase__ : '''simple docstring''' def __init__( self : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : str=13 , _UpperCAmelCase : List[str]=3 , _UpperCAmelCase : List[Any]=32 , _UpperCAmelCase : Any=0.25 , _UpperCAmelCase : Optional[int]=8 , _UpperCAmelCase : Union[str, Any]=8 , _UpperCAmelCase : Tuple=6 , _UpperCAmelCase : Dict=32 , _UpperCAmelCase : Dict=True , _UpperCAmelCase : Optional[Any]=True , _UpperCAmelCase : str=True , _UpperCAmelCase : int="relu6" , _UpperCAmelCase : Optional[int]=1280 , _UpperCAmelCase : str=0.1 , _UpperCAmelCase : List[str]=0.02 , _UpperCAmelCase : Optional[int]=True , _UpperCAmelCase : str=True , _UpperCAmelCase : Optional[int]=10 , _UpperCAmelCase : Optional[Any]=None , ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = image_size UpperCAmelCase_ = depth_multiplier UpperCAmelCase_ = depth_divisible_by UpperCAmelCase_ = min_depth UpperCAmelCase_ = expand_ratio UpperCAmelCase_ = tf_padding UpperCAmelCase_ = output_stride UpperCAmelCase_ = first_layer_is_expansion UpperCAmelCase_ = finegrained_output UpperCAmelCase_ = hidden_act UpperCAmelCase_ = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) UpperCAmelCase_ = classifier_dropout_prob UpperCAmelCase_ = use_labels UpperCAmelCase_ = is_training UpperCAmelCase_ = num_labels UpperCAmelCase_ = initializer_range UpperCAmelCase_ = scope def lowercase__ ( self : Optional[Any] ) -> int: '''simple docstring''' UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels, pixel_labels def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowercase__ ( self : int , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = MobileNetVaModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def lowercase__ ( self : Dict , _UpperCAmelCase : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : int ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = MobileNetVaForImageClassification(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int , _UpperCAmelCase : Dict , _UpperCAmelCase : int ) -> str: '''simple docstring''' UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = MobileNetVaForSemanticSegmentation(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() UpperCAmelCase_ = model(_UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) UpperCAmelCase_ = model(_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowercase__ ( self : Dict ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowercase__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) UpperCamelCase = ( { '''feature-extraction''': MobileNetVaModel, '''image-classification''': MobileNetVaForImageClassification, '''image-segmentation''': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def lowercase__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = MobileNetVaModelTester(self ) UpperCAmelCase_ = MobileNetVaConfigTester(self , config_class=_UpperCAmelCase , has_text_modality=_UpperCAmelCase ) def lowercase__ ( self : Optional[Any] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileNetV2 does not use inputs_embeds" ) def lowercase__ ( self : Dict ) -> List[str]: '''simple docstring''' pass @unittest.skip(reason="MobileNetV2 does not support input and output embeddings" ) def lowercase__ ( self : str ) -> int: '''simple docstring''' pass @unittest.skip(reason="MobileNetV2 does not output attentions" ) def lowercase__ ( self : List[Any] ) -> List[str]: '''simple docstring''' pass def lowercase__ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(_UpperCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _UpperCAmelCase ) def lowercase__ ( self : Dict ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def lowercase__ ( self : Any ) -> List[str]: '''simple docstring''' def check_hidden_states_output(_UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int ): UpperCAmelCase_ = model_class(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = 16 self.assertEqual(len(_UpperCAmelCase ) , _UpperCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = True check_hidden_states_output(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True check_hidden_states_output(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def lowercase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_UpperCAmelCase ) def lowercase__ ( self : Optional[int] ) -> Any: '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_UpperCAmelCase ) @slow def lowercase__ ( self : List[Any] ) -> int: '''simple docstring''' for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = MobileNetVaModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def a__ ( ): UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowercase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase__ ( self : Any ) -> Dict: '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained("google/mobilenet_v2_1.0_224" ) if is_vision_available() else None ) @slow def lowercase__ ( self : Dict ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = MobileNetVaForImageClassification.from_pretrained("google/mobilenet_v2_1.0_224" ).to(_UpperCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=_UpperCAmelCase , return_tensors="pt" ).to(_UpperCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**_UpperCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , _UpperCAmelCase ) UpperCAmelCase_ = torch.tensor([0.2445, -1.1993, 0.1905] ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _UpperCAmelCase , atol=1e-4 ) ) @slow def lowercase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = MobileNetVaForSemanticSegmentation.from_pretrained("google/deeplabv3_mobilenet_v2_1.0_513" ) UpperCAmelCase_ = model.to(_UpperCAmelCase ) UpperCAmelCase_ = MobileNetVaImageProcessor.from_pretrained("google/deeplabv3_mobilenet_v2_1.0_513" ) UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=_UpperCAmelCase , return_tensors="pt" ).to(_UpperCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**_UpperCAmelCase ) UpperCAmelCase_ = outputs.logits # verify the logits UpperCAmelCase_ = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , _UpperCAmelCase ) UpperCAmelCase_ = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=_UpperCAmelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , _UpperCAmelCase , atol=1e-4 ) )
82
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch UpperCAmelCase = logging.get_logger(__name__) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : str = ["pixel_values"] def __init__( self : List[Any] , A__ : bool = True , A__ : Optional[Dict[str, int]] = None , A__ : PILImageResampling = PILImageResampling.BILINEAR , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , **A__ : int , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[int] = size if size is not None else {"shortest_edge": 2_56} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[str] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Any = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : int = do_resize snake_case_ : Optional[Any] = size snake_case_ : Optional[Any] = resample snake_case_ : Optional[int] = do_center_crop snake_case_ : List[Any] = crop_size snake_case_ : List[Any] = do_rescale snake_case_ : Optional[int] = rescale_factor snake_case_ : Optional[Any] = do_normalize snake_case_ : List[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : str , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : Any = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : int , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Union[str, Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Tuple = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : float , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Tuple ) -> np.ndarray: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Dict , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , A__ : ImageInput , A__ : Optional[bool] = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : Dict[str, int] = None , A__ : Optional[bool] = None , A__ : Optional[float] = None , A__ : Optional[bool] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **A__ : Union[str, Any] , ) -> Optional[int]: '''simple docstring''' snake_case_ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Dict = size if size is not None else self.size snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : str = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : Dict = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Any = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : Dict = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Tuple = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Any = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : List[str] = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : Any = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : Union[str, Any] = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : Optional[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Any = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : Dict , A__ : List[Tuple] = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A__ ) != len(A__ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(A__ ): snake_case_ : Dict = target_sizes.numpy() snake_case_ : int = [] for idx in range(len(A__ ) ): snake_case_ : List[str] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=A__ ) snake_case_ : int = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A__ ) else: snake_case_ : List[Any] = logits.argmax(dim=1 ) snake_case_ : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
666
0
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def snake_case_ ( A_ : int ): '''simple docstring''' random.seed(A_ ) np.random.seed(A_ ) torch.manual_seed(A_ ) torch.cuda.manual_seed_all(A_ ) # ^^ safe to call this function even if cuda is not available class __snake_case : def __init__( self : int , __lowerCAmelCase : Iterable[torch.nn.Parameter] , __lowerCAmelCase : float = 0.99_99 , __lowerCAmelCase : float = 0.0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : bool = False , __lowerCAmelCase : Union[float, int] = 1.0 , __lowerCAmelCase : Union[float, int] = 2 / 3 , __lowerCAmelCase : Optional[Any] = None , __lowerCAmelCase : Dict[str, Any] = None , **__lowerCAmelCase : Optional[Any] , ): """simple docstring""" if isinstance(__lowerCAmelCase , torch.nn.Module ): _lowerCamelCase : Dict = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage`''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase , ) _lowerCamelCase : int = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _lowerCamelCase : Optional[int] = True if kwargs.get('''max_value''' , __lowerCAmelCase ) is not None: _lowerCamelCase : str = '''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('''max_value''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) _lowerCamelCase : Any = kwargs['''max_value'''] if kwargs.get('''min_value''' , __lowerCAmelCase ) is not None: _lowerCamelCase : Optional[int] = '''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('''min_value''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = kwargs['''min_value'''] _lowerCamelCase : int = list(__lowerCAmelCase ) _lowerCamelCase : int = [p.clone().detach() for p in parameters] if kwargs.get('''device''' , __lowerCAmelCase ) is not None: _lowerCamelCase : Tuple = '''The `device` argument is deprecated. Please use `to` instead.''' deprecate('''device''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase ) self.to(device=kwargs['''device'''] ) _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : Tuple = decay _lowerCamelCase : Any = min_decay _lowerCamelCase : str = update_after_step _lowerCamelCase : Any = use_ema_warmup _lowerCamelCase : Optional[Any] = inv_gamma _lowerCamelCase : Tuple = power _lowerCamelCase : Any = 0 _lowerCamelCase : Any = None # set in `step()` _lowerCamelCase : int = model_cls _lowerCamelCase : Dict = model_config @classmethod def SCREAMING_SNAKE_CASE ( cls : int , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Optional[int] = model_cls.load_config(__lowerCAmelCase , return_unused_kwargs=__lowerCAmelCase ) _lowerCamelCase : int = model_cls.from_pretrained(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = cls(model.parameters() , model_cls=__lowerCAmelCase , model_config=model.config ) ema_model.load_state_dict(__lowerCAmelCase ) return ema_model def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : Optional[int] ): """simple docstring""" if self.model_cls is None: raise ValueError('''`save_pretrained` can only be used if `model_cls` was defined at __init__.''' ) if self.model_config is None: raise ValueError('''`save_pretrained` can only be used if `model_config` was defined at __init__.''' ) _lowerCamelCase : Dict = self.model_cls.from_config(self.model_config ) _lowerCamelCase : int = self.state_dict() state_dict.pop('''shadow_params''' , __lowerCAmelCase ) model.register_to_config(**__lowerCAmelCase ) self.copy_to(model.parameters() ) model.save_pretrained(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[str] = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _lowerCamelCase : List[Any] = 1 - (1 + step / self.inv_gamma) ** -self.power else: _lowerCamelCase : str = (1 + step) / (1_0 + step) _lowerCamelCase : List[Any] = min(__lowerCAmelCase , self.decay ) # make sure decay is not smaller than min_decay _lowerCamelCase : str = max(__lowerCAmelCase , self.min_decay ) return cur_decay_value @torch.no_grad() def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" if isinstance(__lowerCAmelCase , torch.nn.Module ): _lowerCamelCase : int = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( '''passing a `torch.nn.Module` to `ExponentialMovingAverage.step`''' , '''1.0.0''' , __lowerCAmelCase , standard_warn=__lowerCAmelCase , ) _lowerCamelCase : Dict = parameters.parameters() _lowerCamelCase : List[Any] = list(__lowerCAmelCase ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _lowerCamelCase : Optional[int] = self.get_decay(self.optimization_step ) _lowerCamelCase : Optional[Any] = decay _lowerCamelCase : Union[str, Any] = 1 - decay _lowerCamelCase : Dict = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , __lowerCAmelCase ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _lowerCamelCase : Optional[int] = deepspeed.zero.GatheredParameters(__lowerCAmelCase , modifier_rank=__lowerCAmelCase ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" _lowerCamelCase : List[str] = list(__lowerCAmelCase ) for s_param, param in zip(self.shadow_params , __lowerCAmelCase ): param.data.copy_(s_param.to(param.device ).data ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : List[Any]=None , __lowerCAmelCase : List[str]=None ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ p.to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ) if p.is_floating_point() else p.to(device=__lowerCAmelCase ) for p in self.shadow_params ] def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def SCREAMING_SNAKE_CASE ( self : Tuple , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" _lowerCamelCase : Optional[Any] = [param.detach().cpu().clone() for param in parameters] def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __lowerCAmelCase : Iterable[torch.nn.Parameter] ): """simple docstring""" if self.temp_stored_params is None: raise RuntimeError('''This ExponentialMovingAverage has no `store()`ed weights ''' '''to `restore()`''' ) for c_param, param in zip(self.temp_stored_params , __lowerCAmelCase ): param.data.copy_(c_param.data ) # Better memory-wise. _lowerCamelCase : str = None def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : dict ): """simple docstring""" _lowerCamelCase : Any = copy.deepcopy(__lowerCAmelCase ) _lowerCamelCase : str = state_dict.get('''decay''' , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('''Decay must be between 0 and 1''' ) _lowerCamelCase : Any = state_dict.get('''min_decay''' , self.min_decay ) if not isinstance(self.min_decay , __lowerCAmelCase ): raise ValueError('''Invalid min_decay''' ) _lowerCamelCase : Union[str, Any] = state_dict.get('''optimization_step''' , self.optimization_step ) if not isinstance(self.optimization_step , __lowerCAmelCase ): raise ValueError('''Invalid optimization_step''' ) _lowerCamelCase : Optional[Any] = state_dict.get('''update_after_step''' , self.update_after_step ) if not isinstance(self.update_after_step , __lowerCAmelCase ): raise ValueError('''Invalid update_after_step''' ) _lowerCamelCase : Tuple = state_dict.get('''use_ema_warmup''' , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , __lowerCAmelCase ): raise ValueError('''Invalid use_ema_warmup''' ) _lowerCamelCase : int = state_dict.get('''inv_gamma''' , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError('''Invalid inv_gamma''' ) _lowerCamelCase : Tuple = state_dict.get('''power''' , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError('''Invalid power''' ) _lowerCamelCase : List[Any] = state_dict.get('''shadow_params''' , __lowerCAmelCase ) if shadow_params is not None: _lowerCamelCase : Optional[Any] = shadow_params if not isinstance(self.shadow_params , __lowerCAmelCase ): raise ValueError('''shadow_params must be a list''' ) if not all(isinstance(__lowerCAmelCase , torch.Tensor ) for p in self.shadow_params ): raise ValueError('''shadow_params must all be Tensors''' )
83
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) return n == n[::-1] def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 100_0000 ): lowercase = 0 for i in range(1 , __SCREAMING_SNAKE_CASE ): if is_palindrome(__SCREAMING_SNAKE_CASE ) and is_palindrome(bin(__SCREAMING_SNAKE_CASE ).split('b' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
84
from ...configuration_utils import PretrainedConfig UpperCAmelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "tapas" def __init__( self : List[Any] , A__ : str=3_05_22 , A__ : Tuple=7_68 , A__ : List[Any]=12 , A__ : Optional[Any]=12 , A__ : Union[str, Any]=30_72 , A__ : Dict="gelu" , A__ : List[Any]=0.1 , A__ : str=0.1 , A__ : List[Any]=10_24 , A__ : Optional[int]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , A__ : Union[str, Any]=0.02 , A__ : Tuple=1E-12 , A__ : Tuple=0 , A__ : Any=10.0 , A__ : List[str]=0 , A__ : List[str]=1.0 , A__ : Optional[Any]=None , A__ : Tuple=1.0 , A__ : Union[str, Any]=False , A__ : Any=None , A__ : Union[str, Any]=1.0 , A__ : int=1.0 , A__ : str=False , A__ : int=False , A__ : Optional[Any]="ratio" , A__ : str=None , A__ : int=None , A__ : Dict=64 , A__ : int=32 , A__ : Optional[Any]=False , A__ : List[str]=True , A__ : List[Any]=False , A__ : str=False , A__ : Any=True , A__ : Tuple=False , A__ : str=None , A__ : str=None , **A__ : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=A__ , **A__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) snake_case_ : int = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[int] = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : List[Any] = type_vocab_sizes snake_case_ : str = initializer_range snake_case_ : Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters snake_case_ : Optional[int] = positive_label_weight snake_case_ : Dict = num_aggregation_labels snake_case_ : List[str] = aggregation_loss_weight snake_case_ : str = use_answer_as_supervision snake_case_ : int = answer_loss_importance snake_case_ : Any = use_normalized_answer_loss snake_case_ : int = huber_loss_delta snake_case_ : List[Any] = temperature snake_case_ : str = aggregation_temperature snake_case_ : List[str] = use_gumbel_for_cells snake_case_ : List[str] = use_gumbel_for_aggregation snake_case_ : Dict = average_approximation_function snake_case_ : List[str] = cell_selection_preference snake_case_ : Dict = answer_loss_cutoff snake_case_ : List[str] = max_num_rows snake_case_ : Union[str, Any] = max_num_columns snake_case_ : str = average_logits_per_cell snake_case_ : Union[str, Any] = select_one_column snake_case_ : Dict = allow_empty_column_selection snake_case_ : List[Any] = init_cell_selection_weights_to_zero snake_case_ : str = reset_position_index_per_cell snake_case_ : List[Any] = disable_per_token_loss # Aggregation hyperparameters snake_case_ : List[str] = aggregation_labels snake_case_ : Union[str, Any] = no_aggregation_label_index if isinstance(self.aggregation_labels , A__ ): snake_case_ : Optional[int] = {int(A__ ): v for k, v in aggregation_labels.items()}
666
0
import os import pytest from attr import dataclass SCREAMING_SNAKE_CASE__ : int = "us-east-1" # defaults region @dataclass class snake_case : lowercase_ = 42 lowercase_ = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' lowercase_ = { 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 500, 'save_steps': 5_500, } lowercase_ = {**hyperparameters, 'max_steps': 1_000} @property def __lowercase( self : List[str] )-> str: """simple docstring""" if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __lowercase( self : Union[str, Any] )-> str: """simple docstring""" return F'''{self.framework}-transfromers-test''' @property def __lowercase( self : int )-> str: """simple docstring""" return F'''./tests/sagemaker/scripts/{self.framework}''' @property def __lowercase( self : Tuple )-> str: """simple docstring""" if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def _a ( lowercase__ : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = SageMakerTestEnvironment(framework=request.cls.framework )
85
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : str , A__ : str ) -> Optional[int]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : int , A__ : Optional[int] , A__ : Dict ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Any , A__ : List[str] , A__ : str ) -> Optional[int]: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : List[Any] , A__ : List[str] , A__ : Optional[int] ) -> List[str]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class snake_case__ ( _UpperCamelCase ): @require_beam def UpperCAmelCase__ ( self : str ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Dict = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[int] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : str ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam snake_case_ : Tuple = beam.io.parquetio.WriteToParquet snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[Any] = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: snake_case_ : int = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[Any] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Tuple = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[str] = NestedBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) snake_case_ : int = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
666
0
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def __snake_case ( __UpperCamelCase : str ,__UpperCamelCase : str ,**__UpperCamelCase : str ): """simple docstring""" A_ = AutoConfig.from_pretrained(__UpperCamelCase ,**__UpperCamelCase ) A_ = AutoModelForSeqaSeqLM.from_config(__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) AutoTokenizer.from_pretrained(__UpperCamelCase ).save_pretrained(__UpperCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
86
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple ): class snake_case__ : def __init__( self : Any , A__ : List[str] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = metric_id class snake_case__ : _SCREAMING_SNAKE_CASE : List[str] = [MetricMock(_UpperCamelCase ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: int , lowerCAmelCase_: List[Any] , lowerCAmelCase_: Any , lowerCAmelCase_: List[str] ): if "tmp_path" in args: snake_case_ : List[Any] = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowerCAmelCase_ , match="https://huggingface.co/docs/evaluate" ): func(*lowerCAmelCase_ )
666
0
from __future__ import annotations from math import pow, sqrt def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> dict[str, float]: """simple docstring""" if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
87
from __future__ import annotations import bisect def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Any = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: snake_case_ : Tuple = mid + 1 else: snake_case_ : Dict = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Optional[Any] = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: snake_case_ : Optional[Any] = mid + 1 else: snake_case_ : Tuple = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_left(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_right(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Dict = 0 snake_case_ : Tuple = len(lowerCAmelCase_ ) - 1 while left <= right: snake_case_ : int = left + (right - left) // 2 snake_case_ : Optional[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: snake_case_ : Optional[Any] = midpoint - 1 else: snake_case_ : Optional[int] = midpoint + 1 return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Optional[int] = bisect.bisect_left(lowerCAmelCase_ , lowerCAmelCase_ ) if index != len(lowerCAmelCase_ ) and sorted_collection[index] == item: return index return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int , lowerCAmelCase_: int ): if right < left: return None snake_case_ : List[Any] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , midpoint - 1 ) else: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , midpoint + 1 , lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = input("Enter numbers separated by comma:\n").strip() UpperCAmelCase = sorted(int(item) for item in user_input.split(",")) UpperCAmelCase = int(input("Enter a single number to be found in the list:\n")) UpperCAmelCase = binary_search(collection, target) if result is None: print(F"{target} was not found in {collection}.") else: print(F"{target} was found at position {result} in {collection}.")
666
0
"""simple docstring""" from math import sqrt def _snake_case ( __snake_case : int ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(sqrt(__snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _snake_case ( __snake_case : int = 10001 ): """simple docstring""" _lowerCamelCase : str = 0 _lowerCamelCase : Any = 1 while count != nth and number < 3: number += 1 if is_prime(__snake_case ): count += 1 while count != nth: number += 2 if is_prime(__snake_case ): count += 1 return number if __name__ == "__main__": print(f'''{solution() = }''')
88
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , A__ : VQModel , A__ : UNetaDModel , A__ : DDIMScheduler ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , A__ : int = 1 , A__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A__ : float = 0.0 , A__ : int = 50 , A__ : Optional[str] = "pil" , A__ : bool = True , **A__ : Optional[Any] , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' snake_case_ : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) snake_case_ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler snake_case_ : Any = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature snake_case_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case_ : List[Any] = {} if accepts_eta: snake_case_ : int = eta for t in self.progress_bar(self.scheduler.timesteps ): snake_case_ : Union[str, Any] = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual snake_case_ : Dict = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case_ : Union[str, Any] = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE snake_case_ : int = self.vqvae.decode(A__ ).sample snake_case_ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ : Optional[int] = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
666
0
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> list: if len(lowerCamelCase_ ) == 0: return [] _lowercase , _lowercase : Any = min(lowerCamelCase_ ), max(lowerCamelCase_ ) _lowercase : Tuple = int(max_value - min_value ) + 1 _lowercase : list[list] = [[] for _ in range(lowerCamelCase_ )] for i in my_list: buckets[int(i - min_value )].append(lowerCamelCase_ ) return [v for bucket in buckets for v in sorted(lowerCamelCase_ )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
89
from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) snake_case_ : List[str] = precision snake_case_ : Union[str, Any] = ceil(precision / 1_4 ) snake_case_ : List[str] = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : str = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : str = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): snake_case_ : Tuple = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCAmelCase = 5_0 print(F"The first {n} digits of pi is: {pi(n)}")
666
0
'''simple docstring''' def _snake_case ( A , A , A ) -> float: return round(float(moles / volume ) * nfactor ) def _snake_case ( A , A , A ) -> float: return round(float((moles * 0.0_821 * temperature) / (volume) ) ) def _snake_case ( A , A , A ) -> float: return round(float((moles * 0.0_821 * temperature) / (pressure) ) ) def _snake_case ( A , A , A ) -> float: return round(float((pressure * volume) / (0.0_821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
90
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int = 1_0_0_0 ): snake_case_ ,snake_case_ : List[str] = 1, 1 snake_case_ : List[str] = 2 while True: snake_case_ : Tuple = 0 snake_case_ : Union[str, Any] = fa + fa snake_case_ ,snake_case_ : str = fa, f index += 1 for _ in str(lowerCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
666
0
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[str] ,A_ : Tuple ,A_ : List[str]=2 ,A_ : Optional[Any]=32 ,A_ : Union[str, Any]=16 ,A_ : Dict=3 ,A_ : Optional[Any]=True ,A_ : Optional[int]=True ,A_ : Optional[int]=32 ,A_ : List[Any]=4 ,A_ : Dict=[0, 1, 2, 3] ,A_ : str=4 ,A_ : List[Any]=37 ,A_ : Dict="gelu" ,A_ : Optional[Any]=0.1 ,A_ : Tuple=0.1 ,A_ : List[Any]=0.02 ,A_ : List[str]=3 ,A_ : Tuple=[1, 384, 24, 24] ,A_ : List[str]=True ,A_ : Union[str, Any]=None ,) -> Any: A = parent A = batch_size A = image_size A = patch_size A = num_channels A = is_training A = use_labels A = hidden_size A = num_hidden_layers A = backbone_out_indices A = num_attention_heads A = intermediate_size A = hidden_act A = hidden_dropout_prob A = attention_probs_dropout_prob A = initializer_range A = num_labels A = backbone_featmap_shape A = scope A = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) A = (image_size // patch_size) ** 2 A = num_patches + 1 def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Tuple: A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A = None if self.use_labels: A = ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) A = self.get_config() return config, pixel_values, labels def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: A = { 'global_padding': 'same', 'layer_type': 'bottleneck', 'depths': [3, 4, 9], 'out_features': ['stage1', 'stage2', 'stage3'], 'embedding_dynamic_padding': True, 'hidden_sizes': [96, 192, 384, 768], 'num_groups': 2, } return DPTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,backbone_out_indices=self.backbone_out_indices ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=A_ ,initializer_range=self.initializer_range ,is_hybrid=self.is_hybrid ,backbone_config=A_ ,backbone_featmap_shape=self.backbone_featmap_shape ,) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ,A_ : Dict ,A_ : Optional[Any] ,A_ : Dict ) -> List[str]: A = DPTModel(config=A_ ) model.to(A_ ) model.eval() A = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _SCREAMING_SNAKE_CASE ( self : Tuple ,A_ : Any ,A_ : int ,A_ : Tuple ) -> Dict: A = self.num_labels A = DPTForDepthEstimation(A_ ) model.to(A_ ) model.eval() A = model(A_ ) self.parent.assertEqual(result.predicted_depth.shape ,(self.batch_size, self.image_size, self.image_size) ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ,A_ : List[Any] ,A_ : str ,A_ : int ) -> List[str]: A = self.num_labels A = DPTForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() A = model(A_ ,labels=A_ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size, self.image_size) ) def _SCREAMING_SNAKE_CASE ( self : int ) -> int: A = self.prepare_config_and_inputs() A , A , A = config_and_inputs A = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( _lowercase , _lowercase , unittest.TestCase ): '''simple docstring''' _lowerCamelCase: Union[str, Any] = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () _lowerCamelCase: Optional[Any] = ( { '''depth-estimation''': DPTForDepthEstimation, '''feature-extraction''': DPTModel, '''image-segmentation''': DPTForSemanticSegmentation, } if is_torch_available() else {} ) _lowerCamelCase: Union[str, Any] = False _lowerCamelCase: Dict = False _lowerCamelCase: str = False def _SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: A = DPTModelTester(self ) A = ConfigTester(self ,config_class=A_ ,has_text_modality=A_ ,hidden_size=37 ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> Any: self.config_tester.run_common_tests() @unittest.skip(reason='DPT does not use inputs_embeds' ) def _SCREAMING_SNAKE_CASE ( self : int ) -> Any: pass def _SCREAMING_SNAKE_CASE ( self : str ) -> Dict: A , A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) A = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ ,nn.Linear ) ) def _SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: A , A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A = model_class(A_ ) A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A = [*signature.parameters.keys()] A = ['pixel_values'] self.assertListEqual(arg_names[:1] ,A_ ) def _SCREAMING_SNAKE_CASE ( self : int ) -> int: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> int: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[Any]: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> int: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue A , A = self.model_tester.prepare_config_and_inputs_for_common() A = True if model_class in get_values(A_ ): continue A = model_class(A_ ) model.to(A_ ) model.train() A = self._prepare_for_class(A_ ,A_ ,return_labels=A_ ) A = model(**A_ ).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue A , A = self.model_tester.prepare_config_and_inputs_for_common() A = False A = True if model_class in get_values(A_ ) or not model_class.supports_gradient_checkpointing: continue A = model_class(A_ ) model.to(A_ ) model.gradient_checkpointing_enable() model.train() A = self._prepare_for_class(A_ ,A_ ,return_labels=A_ ) A = model(**A_ ).loss loss.backward() def _SCREAMING_SNAKE_CASE ( self : int ) -> Tuple: A , A = self.model_tester.prepare_config_and_inputs_for_common() A = _config_zero_init(A_ ) for model_class in self.all_model_classes: A = model_class(config=A_ ) # Skip the check for the backbone A = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": A = [F'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=F'Parameter {name} of model {model_class} seems not properly initialized' ,) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Tuple: pass @slow def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: A = DPTModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type A , A = self.model_tester.prepare_config_and_inputs_for_common() A = 'add' with self.assertRaises(A_ ): A = DPTForDepthEstimation(A_ ) def _snake_case ( ): A = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision @slow class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Dict: A = DPTImageProcessor.from_pretrained('Intel/dpt-hybrid-midas' ) A = DPTForDepthEstimation.from_pretrained('Intel/dpt-hybrid-midas' ).to(A_ ) A = prepare_img() A = image_processor(images=A_ ,return_tensors='pt' ).to(A_ ) # forward pass with torch.no_grad(): A = model(**A_ ) A = outputs.predicted_depth # verify the predicted depth A = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape ,A_ ) A = torch.tensor( [[[5.64_37, 5.61_46, 5.65_11], [5.43_71, 5.56_49, 5.59_58], [5.52_15, 5.51_84, 5.52_93]]] ).to(A_ ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 ,A_ ,atol=1e-4 ) )
91
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int | float] , lowerCAmelCase_: int , lowerCAmelCase_: int ): if len(lowerCAmelCase_ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(lowerCAmelCase_ ) or left < -len(lowerCAmelCase_ ) or right >= len(lowerCAmelCase_ ) or right < -len(lowerCAmelCase_ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case_ : List[Any] = (left + right) >> 1 # the middle snake_case_ : Dict = find_max(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # find max in range[left, mid] snake_case_ : int = find_max(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
666
0
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def _lowerCAmelCase ( __magic_name__ : Dict , __magic_name__ : Dict , __magic_name__ : Optional[Any]=1024 , __magic_name__ : Dict=1024 , __magic_name__ : Dict=False , **__magic_name__ : List[Any] ) -> str: lowercase : int =AutoTokenizer.from_pretrained(__magic_name__ ) lowercase : int =SeqaSeqDataset(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , type_path='''train''' , **__magic_name__ ) lowercase : List[Any] =tok.pad_token_id def get_lens(__magic_name__ : List[Any] ): lowercase : str =tqdm( DataLoader(__magic_name__ , batch_size=512 , num_workers=8 , shuffle=__magic_name__ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) lowercase : Optional[int] =[] for batch in dl: lowercase : Optional[Any] =batch['''input_ids'''].ne(__magic_name__ ).sum(1 ).tolist() lowercase : Optional[int] =batch['''labels'''].ne(__magic_name__ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__magic_name__ , __magic_name__ ): max_lens.append(max(__magic_name__ , __magic_name__ ) ) else: max_lens.extend(__magic_name__ ) return max_lens lowercase : int =get_lens(__magic_name__ ) lowercase : List[str] =SeqaSeqDataset(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , type_path='''val''' , **__magic_name__ ) lowercase : Dict =get_lens(__magic_name__ ) pickle_save(__magic_name__ , train_ds.len_file ) pickle_save(__magic_name__ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
92
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } UpperCAmelCase = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE : List[str] = RobertaTokenizer def __init__( self : Optional[int] , A__ : List[Any]=None , A__ : Optional[int]=None , A__ : List[str]=None , A__ : Dict="replace" , A__ : List[str]="<s>" , A__ : Optional[Any]="</s>" , A__ : List[str]="</s>" , A__ : List[Any]="<s>" , A__ : int="<unk>" , A__ : int="<pad>" , A__ : List[Any]="<mask>" , A__ : Any=False , A__ : Optional[int]=True , **A__ : Union[str, Any] , ) -> int: '''simple docstring''' super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : List[Any] = getattr(A__ , pre_tok_state.pop("type" ) ) snake_case_ : Any = add_prefix_space snake_case_ : List[Any] = pre_tok_class(**A__ ) snake_case_ : Optional[int] = add_prefix_space snake_case_ : List[str] = "post_processor" snake_case_ : Tuple = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case_ : str = tuple(state["sep"] ) if "cls" in state: snake_case_ : Tuple = tuple(state["cls"] ) snake_case_ : Tuple = False if state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : Optional[Any] = add_prefix_space snake_case_ : str = True if state.get("trim_offsets" , A__ ) != trim_offsets: snake_case_ : Optional[int] = trim_offsets snake_case_ : List[Any] = True if changes_to_apply: snake_case_ : int = getattr(A__ , state.pop("type" ) ) snake_case_ : List[Any] = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCAmelCase__ ( self : Optional[Any] ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCAmelCase__ ( self : Tuple , A__ : Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case_ : Any = value def UpperCAmelCase__ ( self : int , *A__ : Optional[Any] , **A__ : int ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[Any] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , *A__ : Any , **A__ : List[Any] ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[int] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCAmelCase__ ( self : int , A__ : List[str] , A__ : Union[str, Any]=None ) -> Any: '''simple docstring''' snake_case_ : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Dict , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : str = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
666
0
"""simple docstring""" import sys __A = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def __A (_SCREAMING_SNAKE_CASE = N ) ->int: """simple docstring""" lowerCAmelCase__ :Any = -sys.maxsize - 1 for i in range(len(_SCREAMING_SNAKE_CASE ) - 12 ): lowerCAmelCase__ :List[Any] = 1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: lowerCAmelCase__ :List[str] = product return largest_product if __name__ == "__main__": print(F'''{solution() = }''')
93
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) snake_case_ : Any = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } snake_case_ : List[str] = model(A__ )["last_hidden_state"] snake_case_ : str = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , A__ ) # compare the actual values for a slice. snake_case_ : List[str] = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
666
0
'''simple docstring''' from __future__ import annotations SCREAMING_SNAKE_CASE = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class UpperCAmelCase_ : """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase : dict[str, list[str]] , UpperCAmelCase : str ) -> None: '''simple docstring''' lowercase : Any =graph # mapping node to its parent in resulting breadth first tree lowercase : dict[str, str | None] ={} lowercase : int =source_vertex def A__ ( self : Optional[Any] ) -> None: '''simple docstring''' lowercase : str ={self.source_vertex} lowercase : Optional[int] =None lowercase : List[str] =[self.source_vertex] # first in first out queue while queue: lowercase : Optional[int] =queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(UpperCAmelCase ) lowercase : Optional[Any] =vertex queue.append(UpperCAmelCase ) def A__ ( self : Dict , UpperCAmelCase : str ) -> str: '''simple docstring''' if target_vertex == self.source_vertex: return self.source_vertex lowercase : int =self.parent.get(UpperCAmelCase ) if target_vertex_parent is None: lowercase : List[str] =( f'No path from vertex: {self.source_vertex} to vertex: {target_vertex}' ) raise ValueError(UpperCAmelCase ) return self.shortest_path(UpperCAmelCase ) + f'->{target_vertex}' if __name__ == "__main__": SCREAMING_SNAKE_CASE = Graph(graph, 'G') g.breath_first_search() print(g.shortest_path('D')) print(g.shortest_path('G')) print(g.shortest_path('Foo'))
94
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = ["pixel_values"] def __init__( self : Union[str, Any] , A__ : bool = True , A__ : Dict[str, int] = None , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = True , **A__ : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : str = size if size is not None else {"shortest_edge": 2_24} snake_case_ : Union[str, Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : str = size snake_case_ : Optional[Any] = resample snake_case_ : Any = do_center_crop snake_case_ : Any = crop_size snake_case_ : str = do_rescale snake_case_ : Optional[Any] = rescale_factor snake_case_ : int = do_normalize snake_case_ : Optional[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ : List[str] = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ : int = do_convert_rgb def UpperCAmelCase__ ( self : Optional[int] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : str = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : np.ndarray , A__ : Union[int, float] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> str: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Any , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Any , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[Any] , A__ : ImageInput , A__ : bool = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : int = None , A__ : bool = None , A__ : float = None , A__ : bool = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A__ : Optional[Any] , ) -> PIL.Image.Image: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Union[str, Any] = size if size is not None else self.size snake_case_ : Any = get_size_dict(A__ , param_name="size" , default_to_square=A__ ) snake_case_ : Optional[int] = resample if resample is not None else self.resample snake_case_ : int = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" , default_to_square=A__ ) snake_case_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : List[Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ : List[Any] = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ : Dict = [convert_to_rgb(A__ ) for image in images] # All transformations expect numpy arrays. snake_case_ : Dict = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : Tuple = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : int = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : List[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Tuple = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ )
666
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCamelCase_ = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] lowerCamelCase_ = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } lowerCamelCase_ = {f'funnel-transformer/{name}': 512 for name in _model_names} lowerCamelCase_ = {f'funnel-transformer/{name}': {'''do_lower_case''': True} for name in _model_names} class UpperCamelCase_ (__A ): __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_INIT_CONFIGURATION __magic_name__ = FunnelTokenizer __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = 2 def __init__( self : Union[str, Any] , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : Optional[int]=None , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Any="<unk>" , lowerCAmelCase_ : Tuple="<sep>" , lowerCAmelCase_ : Tuple="<pad>" , lowerCAmelCase_ : Optional[Any]="<cls>" , lowerCAmelCase_ : Tuple="<mask>" , lowerCAmelCase_ : int="<s>" , lowerCAmelCase_ : Dict="</s>" , lowerCAmelCase_ : Union[str, Any]=True , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : int=None , lowerCAmelCase_ : Union[str, Any]="##" , **lowerCAmelCase_ : str , ) -> int: super().__init__( lowerCAmelCase_ , tokenizer_file=lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , clean_text=lowerCAmelCase_ , tokenize_chinese_chars=lowerCAmelCase_ , strip_accents=lowerCAmelCase_ , wordpieces_prefix=lowerCAmelCase_ , **lowerCAmelCase_ , ) UpperCAmelCase_ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase_ ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase_ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase_ ) != tokenize_chinese_chars ): UpperCAmelCase_ : str = getattr(lowerCAmelCase_ , normalizer_state.pop("type" ) ) UpperCAmelCase_ : List[Any] = do_lower_case UpperCAmelCase_ : List[Any] = strip_accents UpperCAmelCase_ : Dict = tokenize_chinese_chars UpperCAmelCase_ : Optional[Any] = normalizer_class(**lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = do_lower_case def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Optional[int]=None ) -> Optional[Any]: UpperCAmelCase_ : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase_ : Tuple = [self.sep_token_id] UpperCAmelCase_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> Tuple[str]: UpperCAmelCase_ : Union[str, Any] = self._tokenizer.model.save(lowerCAmelCase_ , name=lowerCAmelCase_ ) return tuple(lowerCAmelCase_ )
95
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): snake_case_ ,snake_case_ : Dict = position snake_case_ : int = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] snake_case_ : Union[str, Any] = [] for position in positions: snake_case_ ,snake_case_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(lowerCAmelCase_ ) return permissible_positions def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] , lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): if is_complete(lowerCAmelCase_ ): return True for position in get_valid_pos(lowerCAmelCase_ , len(lowerCAmelCase_ ) ): snake_case_ ,snake_case_ : Dict = position if board[y][x] == 0: snake_case_ : List[str] = curr + 1 if open_knight_tour_helper(lowerCAmelCase_ , lowerCAmelCase_ , curr + 1 ): return True snake_case_ : Dict = 0 return False def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): snake_case_ : Optional[Any] = 1 if open_knight_tour_helper(lowerCAmelCase_ , (i, j) , 1 ): return board snake_case_ : Dict = 0 snake_case_ : str = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
666
0
"""simple docstring""" from collections import namedtuple import requests from lxml import html # type: ignore __lowerCamelCase = namedtuple('covid_data', 'cases deaths recovered') def a ( __UpperCAmelCase : str = "https://www.worldometers.info/coronavirus/" ) -> covid_data: __magic_name__: Dict = """//div[@class = \"maincounter-number\"]/span/text()""" return covid_data(*html.fromstring(requests.get(__UpperCAmelCase ).content ).xpath(__UpperCAmelCase ) ) __lowerCamelCase = 'Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}' print(fmt.format(*covid_stats()))
96
from ...configuration_utils import PretrainedConfig class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = "bert-generation" def __init__( self : Optional[int] , A__ : List[Any]=5_03_58 , A__ : Any=10_24 , A__ : Any=24 , A__ : List[Any]=16 , A__ : List[Any]=40_96 , A__ : int="gelu" , A__ : List[str]=0.1 , A__ : List[str]=0.1 , A__ : str=5_12 , A__ : int=0.02 , A__ : Any=1E-12 , A__ : Optional[Any]=0 , A__ : List[str]=2 , A__ : Optional[int]=1 , A__ : str="absolute" , A__ : Any=True , **A__ : Optional[Any] , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , **A__ ) snake_case_ : str = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Optional[Any] = hidden_act snake_case_ : Tuple = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : Optional[Any] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = position_embedding_type snake_case_ : Dict = use_cache
666
0
from __future__ import annotations def a ( snake_case__: Optional[int] , snake_case__: Optional[int] , snake_case__: Any , snake_case__: Optional[int] ): # noqa: E741 '''simple docstring''' while r - l > 1: lowercase_ = (l + r) // 2 if v[m] >= key: lowercase_ = m else: lowercase_ = m # noqa: E741 return r def a ( snake_case__: list[int] ): '''simple docstring''' if len(snake_case__ ) == 0: return 0 lowercase_ = [0] * len(snake_case__ ) lowercase_ = 1 lowercase_ = v[0] for i in range(1 , len(snake_case__ ) ): if v[i] < tail[0]: lowercase_ = v[i] elif v[i] > tail[length - 1]: lowercase_ = v[i] length += 1 else: lowercase_ = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
97
import math def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [] snake_case_ : List[str] = 2 snake_case_ : Optional[int] = int(math.sqrt(lowerCAmelCase_ ) ) # Size of every segment snake_case_ : str = [True] * (end + 1) snake_case_ : Any = [] while start <= end: if temp[start] is True: in_prime.append(lowerCAmelCase_ ) for i in range(start * start , end + 1 , lowerCAmelCase_ ): snake_case_ : Union[str, Any] = False start += 1 prime += in_prime snake_case_ : Dict = end + 1 snake_case_ : Dict = min(2 * end , lowerCAmelCase_ ) while low <= n: snake_case_ : Any = [True] * (high - low + 1) for each in in_prime: snake_case_ : Optional[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(lowerCAmelCase_ , high + 1 , lowerCAmelCase_ ): snake_case_ : List[Any] = False for j in range(len(lowerCAmelCase_ ) ): if temp[j] is True: prime.append(j + low ) snake_case_ : int = high + 1 snake_case_ : Union[str, Any] = min(high + end , lowerCAmelCase_ ) return prime print(sieve(1_0**6))
666
0
'''simple docstring''' import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever lowercase__ : Optional[int] = logging.getLogger(__name__) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" def __init__( self : Optional[int] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : str=None ) -> int: '''simple docstring''' super().__init__( lowerCAmelCase__ , question_encoder_tokenizer=lowerCAmelCase__ , generator_tokenizer=lowerCAmelCase__ , index=lowerCAmelCase__ , init_retrieval=lowerCAmelCase__ , ) _UpperCamelCase = None def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : int ) -> int: '''simple docstring''' logger.info('''initializing retrieval''' ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info('''dist initialized''' ) # needs to be set manually _UpperCamelCase = self._infer_socket_ifname() # avoid clash with the NCCL port _UpperCamelCase = str(distributed_port + 1 ) _UpperCamelCase = dist.new_group(ranks=lowerCAmelCase__ , backend='''gloo''' ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info('''dist not initialized / main''' ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def snake_case__ ( self : str ) -> Optional[Any]: '''simple docstring''' return dist.get_rank(group=self.process_group ) == 0 def snake_case__ ( self : List[str] , lowerCAmelCase__ : int , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str]=torch.floataa ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = torch.empty(lowerCAmelCase__ , dtype=lowerCAmelCase__ ) dist.scatter(lowerCAmelCase__ , src=0 , scatter_list=lowerCAmelCase__ , group=self.process_group ) return target_tensor def snake_case__ ( self : int ) -> int: '''simple docstring''' _UpperCamelCase = psutil.net_if_addrs() # a hacky way to deal with varying network interface names _UpperCamelCase = next((addr for addr in addrs if addr.startswith('''e''' )) , lowerCAmelCase__ ) return ifname def snake_case__ ( self : Dict , lowerCAmelCase__ : np.ndarray , lowerCAmelCase__ : int ) -> Tuple[np.ndarray, List[dict]]: '''simple docstring''' if not dist.is_initialized(): _UpperCamelCase , _UpperCamelCase = self._main_retrieve(lowerCAmelCase__ , lowerCAmelCase__ ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowerCAmelCase__ ) # distributed training _UpperCamelCase = dist.get_world_size(group=self.process_group ) # gather logic _UpperCamelCase = None if self._is_main(): _UpperCamelCase = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(lowerCAmelCase__ )] dist.gather(torch.tensor(lowerCAmelCase__ ) , dst=0 , gather_list=lowerCAmelCase__ , group=self.process_group ) # scatter logic _UpperCamelCase = question_hidden_states.shape[0] _UpperCamelCase = [] _UpperCamelCase = [] if self._is_main(): assert len(lowerCAmelCase__ ) == world_size _UpperCamelCase , _UpperCamelCase = self._main_retrieve(torch.cat(lowerCAmelCase__ ).numpy() , lowerCAmelCase__ ) _UpperCamelCase , _UpperCamelCase = torch.tensor(lowerCAmelCase__ ), torch.tensor(lowerCAmelCase__ ) _UpperCamelCase = self._chunk_tensor(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = self._chunk_tensor(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = self._scattered(lowerCAmelCase__ , [n_queries, n_docs] , target_type=torch.intaa ) _UpperCamelCase = self._scattered(lowerCAmelCase__ , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(lowerCAmelCase__ )
98
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self : List[str] , A__ : List[Any] , A__ : int=7 , A__ : Union[str, Any]=3 , A__ : List[str]=30 , A__ : Optional[int]=4_00 , A__ : Optional[Any]=True , A__ : Optional[int]=None , A__ : Optional[Any]=True , A__ : Any=[0.5, 0.5, 0.5] , A__ : int=[0.5, 0.5, 0.5] , A__ : Any=True , A__ : int=1 / 2_55 , A__ : List[str]=True , ) -> Dict: '''simple docstring''' snake_case_ : int = size if size is not None else {"shortest_edge": 18, "longest_edge": 13_33} snake_case_ : Any = parent snake_case_ : Optional[int] = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Union[str, Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Tuple = do_resize snake_case_ : Dict = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : List[Any] = image_std snake_case_ : Tuple = do_rescale snake_case_ : Any = rescale_factor snake_case_ : Optional[int] = do_pad def UpperCAmelCase__ ( self : int ) -> List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase__ ( self : Optional[int] , A__ : Optional[int] , A__ : Any=False ) -> Optional[Any]: '''simple docstring''' if not batched: snake_case_ : Any = image_inputs[0] if isinstance(A__ , Image.Image ): snake_case_ ,snake_case_ : Dict = image.size else: snake_case_ ,snake_case_ : int = image.shape[1], image.shape[2] if w < h: snake_case_ : Dict = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[int] = self.size["shortest_edge"] elif w > h: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : str = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : List[Any] = self.size["shortest_edge"] else: snake_case_ : str = [] for image in image_inputs: snake_case_ ,snake_case_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : List[Any] = max(A__ , key=lambda A__ : item[0] )[0] snake_case_ : int = max(A__ , key=lambda A__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( _UpperCamelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[int] = ConditionalDetrImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' snake_case_ : List[str] = ConditionalDetrImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Any ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A__ , "image_mean" ) ) self.assertTrue(hasattr(A__ , "image_std" ) ) self.assertTrue(hasattr(A__ , "do_normalize" ) ) self.assertTrue(hasattr(A__ , "do_resize" ) ) self.assertTrue(hasattr(A__ , "size" ) ) def UpperCAmelCase__ ( self : List[str] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 13_33} ) self.assertEqual(image_processor.do_pad , A__ ) snake_case_ : Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A__ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , A__ ) def UpperCAmelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' pass def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ ) for image in image_inputs: self.assertIsInstance(A__ , Image.Image ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) snake_case_ : int = image_processing(A__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : int ) -> Any: '''simple docstring''' snake_case_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , numpify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , np.ndarray ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[str] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Optional[int] = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Dict = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : Tuple ) -> str: '''simple docstring''' snake_case_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , torchify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , torch.Tensor ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Any = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : int = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def UpperCAmelCase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : Optional[Any] = json.loads(f.read() ) snake_case_ : int = {"image_id": 3_97_69, "annotations": target} # encode them snake_case_ : Optional[int] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) snake_case_ : Any = image_processing(images=A__ , annotations=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : List[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : List[str] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Tuple = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify orig_size snake_case_ : Any = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) ) @slow def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' snake_case_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : Any = json.loads(f.read() ) snake_case_ : Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_97_69, "segments_info": target} snake_case_ : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor(format="coco_panoptic" ) snake_case_ : str = image_processing(images=A__ , annotations=A__ , masks_path=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : int = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Optional[int] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Optional[int] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify masks snake_case_ : Union[str, Any] = 82_28_73 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , A__ ) # verify orig_size snake_case_ : Dict = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : str = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) )
666
0
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class __UpperCAmelCase ( __A , unittest.TestCase ): """simple docstring""" _lowerCamelCase = XLMProphetNetTokenizer _lowerCamelCase = False _lowerCamelCase = True def snake_case_ ( self ): super().setUp() # We have a SentencePiece fixture for testing __a = XLMProphetNetTokenizer(__A , keep_accents=__A ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case_ ( self ): __a = """[PAD]""" __a = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__A ) , __A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__A ) , __A ) def snake_case_ ( self ): __a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(__A ) , 1012 ) def snake_case_ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def snake_case_ ( self ): __a = XLMProphetNetTokenizer(__A , keep_accents=__A ) __a = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __a = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) __a = tokenizer.convert_tokens_to_ids(__A ) self.assertListEqual( __A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __a = tokenizer.convert_ids_to_tokens(__A ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def snake_case_ ( self ): return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def snake_case_ ( self ): __a = """Hello World!""" __a = [35389, 6672, 49, 2] self.assertListEqual(__A , self.big_tokenizer.encode(__A ) ) @slow def snake_case_ ( self ): # fmt: off __a = {"""input_ids""": [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__A , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
99
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "Model type selected in the list: " + ", ".join(_UpperCamelCase )} ) _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) _SCREAMING_SNAKE_CASE : int = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) _SCREAMING_SNAKE_CASE : float = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) _SCREAMING_SNAKE_CASE : int = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = "train" _SCREAMING_SNAKE_CASE : Any = "dev" class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : SquadDataTrainingArguments _SCREAMING_SNAKE_CASE : List[SquadFeatures] _SCREAMING_SNAKE_CASE : Split _SCREAMING_SNAKE_CASE : bool def __init__( self : str , A__ : SquadDataTrainingArguments , A__ : PreTrainedTokenizer , A__ : Optional[int] = None , A__ : Union[str, Split] = Split.train , A__ : Optional[bool] = False , A__ : Optional[str] = None , A__ : Optional[str] = "pt" , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = args snake_case_ : int = is_language_sensitive snake_case_ : int = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(A__ , A__ ): try: snake_case_ : List[str] = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) snake_case_ : Tuple = mode # Load data features from cache or dataset file snake_case_ : Dict = "v2" if args.version_2_with_negative else "v1" snake_case_ : List[Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case_ : List[Any] = cached_features_file + ".lock" with FileLock(A__ ): if os.path.exists(A__ ) and not args.overwrite_cache: snake_case_ : int = time.time() snake_case_ : List[Any] = torch.load(A__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. snake_case_ : Tuple = self.old_features["features"] snake_case_ : List[str] = self.old_features.get("dataset" , A__ ) snake_case_ : Tuple = self.old_features.get("examples" , A__ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" " future run" ) else: if mode == Split.dev: snake_case_ : Tuple = self.processor.get_dev_examples(args.data_dir ) else: snake_case_ : Tuple = self.processor.get_train_examples(args.data_dir ) snake_case_ ,snake_case_ : Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=A__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=A__ , ) snake_case_ : Any = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , A__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self : str ) -> Dict: '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[int] , A__ : Optional[int] ) -> Dict[str, torch.Tensor]: '''simple docstring''' snake_case_ : Any = self.features[i] snake_case_ : Optional[int] = torch.tensor(feature.input_ids , dtype=torch.long ) snake_case_ : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.cls_index , dtype=torch.long ) snake_case_ : str = torch.tensor(feature.p_mask , dtype=torch.float ) snake_case_ : str = torch.tensor(feature.is_impossible , dtype=torch.float ) snake_case_ : Optional[int] = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: snake_case_ : Any = torch.tensor(feature.start_position , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
666
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _A : List[Any] = logging.get_logger(__name__) _A : str = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} _A : int = { """vocab_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/vocab.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/vocab.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/vocab.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/vocab.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/vocab.json""", }, """merges_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/merges.txt""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/merges.txt""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/merges.txt""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/merges.txt""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/merges.txt""", }, """tokenizer_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/tokenizer.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/tokenizer.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/tokenizer.json""", }, } _A : Optional[Any] = { """gpt2""": 10_24, """gpt2-medium""": 10_24, """gpt2-large""": 10_24, """gpt2-xl""": 10_24, """distilgpt2""": 10_24, } class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase__ : Optional[Any] = VOCAB_FILES_NAMES lowerCamelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["""input_ids""", """attention_mask"""] lowerCamelCase__ : Tuple = GPTaTokenizer def __init__( self , A_=None , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , **A_ , ): '''simple docstring''' super().__init__( A_ , A_ , tokenizer_file=A_ , unk_token=A_ , bos_token=A_ , eos_token=A_ , add_prefix_space=A_ , **A_ , ) SCREAMING_SNAKE_CASE__ = kwargs.pop('''add_bos_token''' , A_ ) SCREAMING_SNAKE_CASE__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , A_ ) != add_prefix_space: SCREAMING_SNAKE_CASE__ = getattr(A_ , pre_tok_state.pop('''type''' ) ) SCREAMING_SNAKE_CASE__ = add_prefix_space SCREAMING_SNAKE_CASE__ = pre_tok_class(**A_ ) SCREAMING_SNAKE_CASE__ = add_prefix_space def lowercase_ ( self , *A_ , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = kwargs.get('''is_split_into_words''' , A_ ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A_ , **A_ ) def lowercase_ ( self , *A_ , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = kwargs.get('''is_split_into_words''' , A_ ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*A_ , **A_ ) def lowercase_ ( self , A_ , A_ = None ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = self._tokenizer.model.save(A_ , name=A_ ) return tuple(A_ ) def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(A_ , add_special_tokens=A_ ) + [self.eos_token_id] ) if len(A_ ) > self.model_max_length: SCREAMING_SNAKE_CASE__ = input_ids[-self.model_max_length :] return input_ids
100
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "git_vision_model" def __init__( self : int , A__ : Union[str, Any]=7_68 , A__ : List[Any]=30_72 , A__ : Tuple=12 , A__ : Optional[Any]=12 , A__ : Optional[int]=3 , A__ : List[str]=2_24 , A__ : Dict=16 , A__ : int="quick_gelu" , A__ : Any=1E-5 , A__ : Tuple=0.0 , A__ : Optional[int]=0.02 , **A__ : List[str] , ) -> Optional[int]: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[Any] = hidden_size snake_case_ : str = intermediate_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = num_channels snake_case_ : Union[str, Any] = patch_size snake_case_ : List[str] = image_size snake_case_ : List[Any] = initializer_range snake_case_ : Any = attention_dropout snake_case_ : Any = layer_norm_eps snake_case_ : int = hidden_act @classmethod def UpperCAmelCase__ ( cls : List[Any] , A__ : Union[str, os.PathLike] , **A__ : Optional[int] ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(A__ ) snake_case_ ,snake_case_ : Tuple = cls.get_config_dict(A__ , **A__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": snake_case_ : Any = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(A__ , **A__ ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = "git" def __init__( self : Any , A__ : List[str]=None , A__ : List[str]=3_05_22 , A__ : Tuple=7_68 , A__ : Tuple=6 , A__ : str=12 , A__ : Any=30_72 , A__ : List[str]="gelu" , A__ : int=0.1 , A__ : Dict=0.1 , A__ : Any=10_24 , A__ : Optional[Any]=0.02 , A__ : Optional[Any]=1E-12 , A__ : Dict=0 , A__ : Any="absolute" , A__ : Tuple=True , A__ : Any=False , A__ : Tuple=1_01 , A__ : Tuple=1_02 , A__ : List[Any]=None , **A__ : List[str] , ) -> int: '''simple docstring''' super().__init__(bos_token_id=A__ , eos_token_id=A__ , pad_token_id=A__ , **A__ ) if vision_config is None: snake_case_ : int = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) snake_case_ : str = GitVisionConfig(**A__ ) snake_case_ : int = vocab_size snake_case_ : List[Any] = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : List[Any] = num_attention_heads snake_case_ : Any = hidden_act snake_case_ : Dict = intermediate_size snake_case_ : Any = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = initializer_range snake_case_ : List[str] = layer_norm_eps snake_case_ : Any = position_embedding_type snake_case_ : Union[str, Any] = use_cache snake_case_ : str = tie_word_embeddings snake_case_ : List[Any] = num_image_with_embedding snake_case_ : Dict = bos_token_id snake_case_ : int = eos_token_id def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' snake_case_ : Tuple = copy.deepcopy(self.__dict__ ) snake_case_ : Optional[int] = self.vision_config.to_dict() snake_case_ : Tuple = self.__class__.model_type return output
666
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ : int =logging.get_logger(__name__) lowerCAmelCase__ : Optional[int] ={ 'google/mobilenet_v2_1.4_224': 'https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json', 'google/mobilenet_v2_1.0_224': 'https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json', 'google/mobilenet_v2_0.75_160': 'https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json', 'google/mobilenet_v2_0.35_96': 'https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = """mobilenet_v2""" def __init__( self , lowerCAmelCase__=3 , lowerCAmelCase__=2_2_4 , lowerCAmelCase__=1.0 , lowerCAmelCase__=8 , lowerCAmelCase__=8 , lowerCAmelCase__=6 , lowerCAmelCase__=3_2 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__="relu6" , lowerCAmelCase__=True , lowerCAmelCase__=0.8 , lowerCAmelCase__=0.02 , lowerCAmelCase__=0.001 , lowerCAmelCase__=2_5_5 , **lowerCAmelCase__ , ): """simple docstring""" super().__init__(**lowerCAmelCase__ ) if depth_multiplier <= 0: raise ValueError('depth_multiplier must be greater than zero.' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = num_channels SCREAMING_SNAKE_CASE_ : int = image_size SCREAMING_SNAKE_CASE_ : int = depth_multiplier SCREAMING_SNAKE_CASE_ : Tuple = depth_divisible_by SCREAMING_SNAKE_CASE_ : int = min_depth SCREAMING_SNAKE_CASE_ : Dict = expand_ratio SCREAMING_SNAKE_CASE_ : Any = output_stride SCREAMING_SNAKE_CASE_ : str = first_layer_is_expansion SCREAMING_SNAKE_CASE_ : Any = finegrained_output SCREAMING_SNAKE_CASE_ : Union[str, Any] = hidden_act SCREAMING_SNAKE_CASE_ : List[str] = tf_padding SCREAMING_SNAKE_CASE_ : Tuple = classifier_dropout_prob SCREAMING_SNAKE_CASE_ : List[Any] = initializer_range SCREAMING_SNAKE_CASE_ : str = layer_norm_eps SCREAMING_SNAKE_CASE_ : Optional[Any] = semantic_loss_ignore_index class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict([('pixel_values', {0: 'batch'})] ) @property def UpperCamelCase__ ( self ): """simple docstring""" if self.task == "image-classification": return OrderedDict([('logits', {0: 'batch'})] ) else: return OrderedDict([('last_hidden_state', {0: 'batch'}), ('pooler_output', {0: 'batch'})] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-4
101
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: str ): def get_matched_characters(lowerCAmelCase_: str , lowerCAmelCase_: str ) -> str: snake_case_ : Tuple = [] snake_case_ : Tuple = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): snake_case_ : str = int(max(0 , i - limit ) ) snake_case_ : Optional[int] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) snake_case_ : List[Any] = f"{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}" return "".join(lowerCAmelCase_ ) # matching characters snake_case_ : List[Any] = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : int = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = len(lowerCAmelCase_ ) # transposition snake_case_ : List[str] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: snake_case_ : str = 0.0 else: snake_case_ : Optional[Any] = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters snake_case_ : Optional[Any] = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("hello", "world"))
666
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ : Union[str, Any] = { """configuration_biogpt""": ["""BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BioGptConfig"""], """tokenization_biogpt""": ["""BioGptTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Union[str, Any] = [ """BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BioGptForCausalLM""", """BioGptForTokenClassification""", """BioGptForSequenceClassification""", """BioGptModel""", """BioGptPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys __magic_name__ : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
102
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) set_seed(7_7_0) UpperCAmelCase = { "c_attn": "att_proj", "c_proj": "out_proj", "c_fc": "in_proj", "transformer.": "", "h.": "layers.", "ln_1": "layernorm_1", "ln_2": "layernorm_2", "ln_f": "layernorm_final", "wpe": "position_embeds_layer", "wte": "input_embeds_layer", } UpperCAmelCase = { "text_small": { "repo_id": "suno/bark", "file_name": "text.pt", }, "coarse_small": { "repo_id": "suno/bark", "file_name": "coarse.pt", }, "fine_small": { "repo_id": "suno/bark", "file_name": "fine.pt", }, "text": { "repo_id": "suno/bark", "file_name": "text_2.pt", }, "coarse": { "repo_id": "suno/bark", "file_name": "coarse_2.pt", }, "fine": { "repo_id": "suno/bark", "file_name": "fine_2.pt", }, } UpperCAmelCase = os.path.dirname(os.path.abspath(__file__)) UpperCAmelCase = os.path.join(os.path.expanduser("~"), ".cache") UpperCAmelCase = os.path.join(os.getenv("XDG_CACHE_HOME", default_cache_dir), "suno", "bark_v0") def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int , lowerCAmelCase_: List[str]=False ): snake_case_ : Union[str, Any] = model_type if use_small: key += "_small" return os.path.join(lowerCAmelCase_ , REMOTE_MODEL_PATHS[key]["file_name"] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: List[str] ): os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) hf_hub_download(repo_id=lowerCAmelCase_ , filename=lowerCAmelCase_ , local_dir=lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Any , lowerCAmelCase_: Dict , lowerCAmelCase_: List[str]=False , lowerCAmelCase_: Dict="text" ): if model_type == "text": snake_case_ : int = BarkSemanticModel snake_case_ : str = BarkSemanticConfig snake_case_ : Optional[Any] = BarkSemanticGenerationConfig elif model_type == "coarse": snake_case_ : str = BarkCoarseModel snake_case_ : Optional[int] = BarkCoarseConfig snake_case_ : Any = BarkCoarseGenerationConfig elif model_type == "fine": snake_case_ : Optional[int] = BarkFineModel snake_case_ : Tuple = BarkFineConfig snake_case_ : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() snake_case_ : Optional[Any] = f"{model_type}_small" if use_small else model_type snake_case_ : Any = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(lowerCAmelCase_ ): logger.info(f"{model_type} model not found, downloading into `{CACHE_DIR}`." ) _download(model_info["repo_id"] , model_info["file_name"] ) snake_case_ : Any = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) # this is a hack snake_case_ : Union[str, Any] = checkpoint["model_args"] if "input_vocab_size" not in model_args: snake_case_ : str = model_args["vocab_size"] snake_case_ : Union[str, Any] = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments snake_case_ : Union[str, Any] = model_args.pop("n_head" ) snake_case_ : int = model_args.pop("n_embd" ) snake_case_ : Any = model_args.pop("n_layer" ) snake_case_ : List[str] = ConfigClass(**checkpoint["model_args"] ) snake_case_ : Optional[Any] = ModelClass(config=lowerCAmelCase_ ) snake_case_ : Tuple = GenerationConfigClass() snake_case_ : List[str] = model_generation_config snake_case_ : Optional[int] = checkpoint["model"] # fixup checkpoint snake_case_ : Optional[int] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(lowerCAmelCase_ ): # replace part of the key with corresponding layer name in HF implementation snake_case_ : Tuple = k[len(lowerCAmelCase_ ) :] for old_layer_name in new_layer_name_dict: snake_case_ : int = new_k.replace(lowerCAmelCase_ , new_layer_name_dict[old_layer_name] ) snake_case_ : int = state_dict.pop(lowerCAmelCase_ ) snake_case_ : Optional[int] = set(state_dict.keys() ) - set(model.state_dict().keys() ) snake_case_ : str = {k for k in extra_keys if not k.endswith(".attn.bias" )} snake_case_ : Any = set(model.state_dict().keys() ) - set(state_dict.keys() ) snake_case_ : List[Any] = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(lowerCAmelCase_ ) != 0: raise ValueError(f"extra keys found: {extra_keys}" ) if len(lowerCAmelCase_ ) != 0: raise ValueError(f"missing keys: {missing_keys}" ) model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) snake_case_ : str = model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) snake_case_ : Union[str, Any] = checkpoint["best_val_loss"].item() logger.info(f"model loaded: {round(n_params/1e6 , 1 )}M params, {round(lowerCAmelCase_ , 3 )} loss" ) model.eval() model.to(lowerCAmelCase_ ) del checkpoint, state_dict return model def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: List[Any] , lowerCAmelCase_: str=False , lowerCAmelCase_: int="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() snake_case_ : int = "cpu" # do conversion on cpu snake_case_ : Optional[Any] = _get_ckpt_path(lowerCAmelCase_ , use_small=lowerCAmelCase_ ) snake_case_ : Tuple = _load_model(lowerCAmelCase_ , lowerCAmelCase_ , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) # load bark initial model snake_case_ : int = _bark_load_model(lowerCAmelCase_ , "cpu" , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) if model_type == "text": snake_case_ : Union[str, Any] = bark_model["model"] if model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model snake_case_ : Optional[Any] = 5 snake_case_ : Optional[int] = 1_0 if model_type in ["text", "coarse"]: snake_case_ : Optional[Any] = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) snake_case_ : str = bark_model(lowerCAmelCase_ )[0] snake_case_ : Tuple = model(lowerCAmelCase_ ) # take last logits snake_case_ : List[str] = output_new_model_total.logits[:, [-1], :] else: snake_case_ : Optional[int] = 3 snake_case_ : str = 8 snake_case_ : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) snake_case_ : Any = model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Union[str, Any] = bark_model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: List[str] , lowerCAmelCase_: Any , lowerCAmelCase_: List[Any] , lowerCAmelCase_: int , lowerCAmelCase_: Optional[Any] , ): snake_case_ : Optional[Any] = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkSemanticConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = BarkCoarseConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[str] = BarkFineConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) snake_case_ : List[str] = BarkSemanticModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkCoarseModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Tuple = BarkFineModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Union[str, Any] = EncodecModel.from_pretrained("facebook/encodec_24khz" ) snake_case_ : Tuple = BarkConfig.from_sub_model_configs( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : List[Any] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) snake_case_ : Optional[int] = BarkModel(lowerCAmelCase_ ) snake_case_ : int = semantic snake_case_ : List[str] = coarseAcoustic snake_case_ : str = fineAcoustic snake_case_ : Optional[Any] = codec snake_case_ : Any = bark_generation_config Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) bark.save_pretrained(lowerCAmelCase_ , repo_id=lowerCAmelCase_ , push_to_hub=lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("model_type", type=str, help="text, coarse or fine.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--is_small", action="store_true", help="convert the small version instead of the large.") UpperCAmelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
666
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class UpperCAmelCase : A__ : str = BlenderbotConfig A__ : Optional[Any] = {} A__ : str = '''gelu''' def __init__( self : Any , __lowerCamelCase : Dict , __lowerCamelCase : int=1_3 , __lowerCamelCase : Union[str, Any]=7 , __lowerCamelCase : Tuple=True , __lowerCamelCase : str=False , __lowerCamelCase : Dict=9_9 , __lowerCamelCase : Optional[Any]=3_2 , __lowerCamelCase : int=2 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : List[Any]=3_7 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Union[str, Any]=2_0 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : str=1 , __lowerCamelCase : str=0 , ): """simple docstring""" _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id def __UpperCAmelCase ( self : Tuple ): """simple docstring""" _snake_case = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _snake_case = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _snake_case = tf.concat([input_ids, eos_tensor] , axis=1 ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _snake_case = prepare_blenderbot_inputs_dict(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return config, inputs_dict def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : str ): """simple docstring""" _snake_case = TFBlenderbotModel(config=__lowerCamelCase ).get_decoder() _snake_case = inputs_dict['''input_ids'''] _snake_case = input_ids[:1, :] _snake_case = inputs_dict['''attention_mask'''][:1, :] _snake_case = inputs_dict['''head_mask'''] _snake_case = 1 # first forward pass _snake_case = model(__lowerCamelCase , attention_mask=__lowerCamelCase , head_mask=__lowerCamelCase , use_cache=__lowerCamelCase ) _snake_case , _snake_case = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _snake_case = tf.concat([input_ids, next_tokens] , axis=-1 ) _snake_case = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _snake_case = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] _snake_case = model(__lowerCamelCase , attention_mask=__lowerCamelCase , past_key_values=__lowerCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _snake_case = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _snake_case = output_from_no_past[:, -3:, random_slice_idx] _snake_case = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowerCamelCase , __lowerCamelCase , rtol=1E-3 ) def snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None , ) -> Union[str, Any]: if attention_mask is None: _snake_case = tf.cast(tf.math.not_equal(lowerCAmelCase_ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _snake_case = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _snake_case = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class UpperCAmelCase ( __SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,unittest.TestCase ): A__ : Dict = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () A__ : Optional[Any] = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () A__ : Dict = ( { '''conversational''': TFBlenderbotForConditionalGeneration, '''feature-extraction''': TFBlenderbotModel, '''summarization''': TFBlenderbotForConditionalGeneration, '''text2text-generation''': TFBlenderbotForConditionalGeneration, '''translation''': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) A__ : str = True A__ : List[Any] = False A__ : Union[str, Any] = False def __UpperCAmelCase ( self : Dict ): """simple docstring""" _snake_case = TFBlenderbotModelTester(self ) _snake_case = ConfigTester(self , config_class=__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Dict ): """simple docstring""" _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowerCamelCase ) @require_tokenizers @require_tf class UpperCAmelCase ( unittest.TestCase ): A__ : List[str] = ['''My friends are cool but they eat too many carbs.'''] A__ : str = '''facebook/blenderbot-400M-distill''' @cached_property def __UpperCAmelCase ( self : Any ): """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def __UpperCAmelCase ( self : Any ): """simple docstring""" _snake_case = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def __UpperCAmelCase ( self : List[str] ): """simple docstring""" _snake_case = self.tokenizer(self.src_text , return_tensors='''tf''' ) _snake_case = self.model.generate( model_inputs.input_ids , ) _snake_case = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__lowerCamelCase )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
103
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
"""simple docstring""" def _lowerCamelCase ( UpperCAmelCase_ : str ) -> Optional[int]: """simple docstring""" A__ , A__ = [], [] while len(UpperCAmelCase_ ) > 1: A__ , A__ = min(UpperCAmelCase_ ), max(UpperCAmelCase_ ) start.append(UpperCAmelCase_ ) end.append(UpperCAmelCase_ ) collection.remove(UpperCAmelCase_ ) collection.remove(UpperCAmelCase_ ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCamelCase = input("""Enter numbers separated by a comma:\n""").strip() UpperCamelCase = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
104
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch UpperCAmelCase = logging.get_logger(__name__) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : str = ["pixel_values"] def __init__( self : List[Any] , A__ : bool = True , A__ : Optional[Dict[str, int]] = None , A__ : PILImageResampling = PILImageResampling.BILINEAR , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , **A__ : int , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[int] = size if size is not None else {"shortest_edge": 2_56} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[str] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Any = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : int = do_resize snake_case_ : Optional[Any] = size snake_case_ : Optional[Any] = resample snake_case_ : Optional[int] = do_center_crop snake_case_ : List[Any] = crop_size snake_case_ : List[Any] = do_rescale snake_case_ : Optional[int] = rescale_factor snake_case_ : Optional[Any] = do_normalize snake_case_ : List[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : str , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : Any = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : int , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Union[str, Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Tuple = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : float , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Tuple ) -> np.ndarray: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Dict , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , A__ : ImageInput , A__ : Optional[bool] = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : Dict[str, int] = None , A__ : Optional[bool] = None , A__ : Optional[float] = None , A__ : Optional[bool] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **A__ : Union[str, Any] , ) -> Optional[int]: '''simple docstring''' snake_case_ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Dict = size if size is not None else self.size snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : str = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : Dict = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Any = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : Dict = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Tuple = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Any = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : List[str] = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : Any = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : Union[str, Any] = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : Optional[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Any = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : Dict , A__ : List[Tuple] = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A__ ) != len(A__ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(A__ ): snake_case_ : Dict = target_sizes.numpy() snake_case_ : int = [] for idx in range(len(A__ ) ): snake_case_ : List[str] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=A__ ) snake_case_ : int = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A__ ) else: snake_case_ : List[Any] = logits.argmax(dim=1 ) snake_case_ : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
666
0
UpperCamelCase__ : List[Any] = '''Alexander Joslin''' import operator as op from .stack import Stack def __UpperCAmelCase ( lowerCamelCase_ : str ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} SCREAMING_SNAKE_CASE_ : Stack[int] = Stack() SCREAMING_SNAKE_CASE_ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(lowerCamelCase_ ) ) elif i in operators: # RULE 2 operator_stack.push(lowerCamelCase_ ) elif i == ")": # RULE 4 SCREAMING_SNAKE_CASE_ : Optional[Any] = operator_stack.peek() operator_stack.pop() SCREAMING_SNAKE_CASE_ : Union[str, Any] = operand_stack.peek() operand_stack.pop() SCREAMING_SNAKE_CASE_ : Tuple = operand_stack.peek() operand_stack.pop() SCREAMING_SNAKE_CASE_ : Any = operators[opr](lowerCamelCase_ , lowerCamelCase_ ) operand_stack.push(lowerCamelCase_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": UpperCamelCase__ : Optional[int] = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
105
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class lowerCAmelCase__ : def __init__( self : Union[str, Any] , __UpperCamelCase : Union[str, Any] , ) -> int: A = parent A = 13 A = 7 A = True A = True A = True A = 99 A = 32 A = 2 A = 4 A = 37 A = 'gelu' A = 0.1 A = 0.1 A = 512 A = 16 A = 2 A = 0.0_2 A = 3 A = 4 A = None def __UpperCamelCase ( self : List[Any] ) -> Any: A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A = None if self.use_input_mask: A = random_attention_mask([self.batch_size, self.seq_length] ) A = None A = None A = None if self.use_labels: A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A = ids_tensor([self.batch_size] , self.num_choices ) A = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase ( self : Tuple ) -> Any: ( ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ) = self.prepare_config_and_inputs() A = True A = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __UpperCamelCase ( self : str , __UpperCamelCase : Optional[Any] , __UpperCamelCase : Dict , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : Optional[Any] ) -> str: A = TFEsmModel(config=__UpperCamelCase ) A = {'input_ids': input_ids, 'attention_mask': input_mask} A = model(__UpperCamelCase ) A = [input_ids, input_mask] A = model(__UpperCamelCase ) A = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase ( self : int , __UpperCamelCase : str , __UpperCamelCase : str , __UpperCamelCase : List[str] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Any , __UpperCamelCase : Tuple , __UpperCamelCase : Any , __UpperCamelCase : str , ) -> Tuple: A = True A = TFEsmModel(config=__UpperCamelCase ) A = { 'input_ids': input_ids, 'attention_mask': input_mask, 'encoder_hidden_states': encoder_hidden_states, 'encoder_attention_mask': encoder_attention_mask, } A = model(__UpperCamelCase ) A = [input_ids, input_mask] A = model(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase ) # Also check the case where encoder outputs are not passed A = model(__UpperCamelCase , attention_mask=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase ( self : str , __UpperCamelCase : int , __UpperCamelCase : Tuple , __UpperCamelCase : Tuple , __UpperCamelCase : Optional[Any] , __UpperCamelCase : str , __UpperCamelCase : Union[str, Any] ) -> Union[str, Any]: A = TFEsmForMaskedLM(config=__UpperCamelCase ) A = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase ( self : int , __UpperCamelCase : Dict , __UpperCamelCase : List[str] , __UpperCamelCase : List[str] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Optional[Any] , __UpperCamelCase : Any ) -> Optional[int]: A = self.num_labels A = TFEsmForTokenClassification(config=__UpperCamelCase ) A = {'input_ids': input_ids, 'attention_mask': input_mask} A = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase ( self : Optional[Any] ) -> List[str]: A = self.prepare_config_and_inputs() ( ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ( A ) , ) = config_and_inputs A = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class lowerCAmelCase__ ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase ): A_ : Tuple = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) A_ : List[Any] = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) A_ : Any = False A_ : str = False def __UpperCamelCase ( self : int ) -> List[str]: A = TFEsmModelTester(self ) A = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __UpperCamelCase ( self : Any ) -> Union[str, Any]: self.config_tester.run_common_tests() def __UpperCamelCase ( self : Union[str, Any] ) -> Union[str, Any]: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __UpperCamelCase ( self : List[str] ) -> List[str]: A = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__UpperCamelCase ) def __UpperCamelCase ( self : List[Any] ) -> int: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __UpperCamelCase ( self : Dict ) -> str: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __UpperCamelCase ( self : Tuple ) -> str: for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A = TFEsmModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip('Protein models do not support embedding resizing.' ) def __UpperCamelCase ( self : int ) -> Union[str, Any]: pass @unittest.skip('Protein models do not support embedding resizing.' ) def __UpperCamelCase ( self : List[Any] ) -> Optional[int]: pass def __UpperCamelCase ( self : Dict ) -> Tuple: A , A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A = model_class(__UpperCamelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer A = model.get_bias() assert isinstance(__UpperCamelCase , __UpperCamelCase ) for k, v in name.items(): assert isinstance(__UpperCamelCase , tf.Variable ) else: A = model.get_output_embeddings() assert x is None A = model.get_bias() assert name is None @require_tf class lowerCAmelCase__ ( unittest.TestCase ): @slow def __UpperCamelCase ( self : Union[str, Any] ) -> Optional[int]: A = TFEsmForMaskedLM.from_pretrained('facebook/esm2_t6_8M_UR50D' ) A = tf.constant([[0, 1, 2, 3, 4, 5]] ) A = model(__UpperCamelCase )[0] A = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , __UpperCamelCase ) # compare the actual values for a slice. A = tf.constant( [ [ [8.9_2_1_5_1_8, -1_0.5_8_9_8_1_4, -6.4_6_7_1_3_0_7], [-6.3_9_6_7_1_5_6, -1_3.9_1_1_3_7_7, -1.1_2_1_1_9_1_5], [-7.7_8_1_2_4_7, -1_3.9_5_1_5_5_7, -3.7_4_0_5_9_2], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-2 ) ) @slow def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: A = TFEsmModel.from_pretrained('facebook/esm2_t6_8M_UR50D' ) A = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) A = model(__UpperCamelCase )[0] # compare the actual values for a slice. A = tf.constant( [ [ [0.1_4_4_4_3_0_9_2, 0.5_4_1_2_5_3_2_7, 0.3_2_4_7_7_3_9], [0.3_0_3_4_0_4_8_4, 0.0_0_5_2_6_6_7_6, 0.3_1_0_7_7_7_2_2], [0.3_2_2_7_8_0_4_3, -0.2_4_9_8_7_0_9_6, 0.3_4_1_4_6_2_8], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
106
from ...configuration_utils import PretrainedConfig UpperCAmelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "tapas" def __init__( self : List[Any] , A__ : str=3_05_22 , A__ : Tuple=7_68 , A__ : List[Any]=12 , A__ : Optional[Any]=12 , A__ : Union[str, Any]=30_72 , A__ : Dict="gelu" , A__ : List[Any]=0.1 , A__ : str=0.1 , A__ : List[Any]=10_24 , A__ : Optional[int]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , A__ : Union[str, Any]=0.02 , A__ : Tuple=1E-12 , A__ : Tuple=0 , A__ : Any=10.0 , A__ : List[str]=0 , A__ : List[str]=1.0 , A__ : Optional[Any]=None , A__ : Tuple=1.0 , A__ : Union[str, Any]=False , A__ : Any=None , A__ : Union[str, Any]=1.0 , A__ : int=1.0 , A__ : str=False , A__ : int=False , A__ : Optional[Any]="ratio" , A__ : str=None , A__ : int=None , A__ : Dict=64 , A__ : int=32 , A__ : Optional[Any]=False , A__ : List[str]=True , A__ : List[Any]=False , A__ : str=False , A__ : Any=True , A__ : Tuple=False , A__ : str=None , A__ : str=None , **A__ : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=A__ , **A__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) snake_case_ : int = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[int] = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : List[Any] = type_vocab_sizes snake_case_ : str = initializer_range snake_case_ : Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters snake_case_ : Optional[int] = positive_label_weight snake_case_ : Dict = num_aggregation_labels snake_case_ : List[str] = aggregation_loss_weight snake_case_ : str = use_answer_as_supervision snake_case_ : int = answer_loss_importance snake_case_ : Any = use_normalized_answer_loss snake_case_ : int = huber_loss_delta snake_case_ : List[Any] = temperature snake_case_ : str = aggregation_temperature snake_case_ : List[str] = use_gumbel_for_cells snake_case_ : List[str] = use_gumbel_for_aggregation snake_case_ : Dict = average_approximation_function snake_case_ : List[str] = cell_selection_preference snake_case_ : Dict = answer_loss_cutoff snake_case_ : List[str] = max_num_rows snake_case_ : Union[str, Any] = max_num_columns snake_case_ : str = average_logits_per_cell snake_case_ : Union[str, Any] = select_one_column snake_case_ : Dict = allow_empty_column_selection snake_case_ : List[Any] = init_cell_selection_weights_to_zero snake_case_ : str = reset_position_index_per_cell snake_case_ : List[Any] = disable_per_token_loss # Aggregation hyperparameters snake_case_ : List[str] = aggregation_labels snake_case_ : Union[str, Any] = no_aggregation_label_index if isinstance(self.aggregation_labels , A__ ): snake_case_ : Optional[int] = {int(A__ ): v for k, v in aggregation_labels.items()}
666
0
'''simple docstring''' import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument( '''--original_config_file''', type=str, required=True, help='''The YAML config file corresponding to the original architecture.''', ) parser.add_argument( '''--num_in_channels''', default=None, type=int, help='''The number of input channels. If `None` number of input channels will be automatically inferred.''', ) parser.add_argument( '''--image_size''', default=5_12, type=int, help=( '''The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2''' ''' Base. Use 768 for Stable Diffusion v2.''' ), ) parser.add_argument( '''--extract_ema''', action='''store_true''', help=( '''Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights''' ''' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield''' ''' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.''' ), ) parser.add_argument( '''--upcast_attention''', action='''store_true''', help=( '''Whether the attention computation should always be upcasted. This is necessary when running stable''' ''' diffusion 2.1.''' ), ) parser.add_argument( '''--from_safetensors''', action='''store_true''', help='''If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.''', ) parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') def _SCREAMING_SNAKE_CASE ( __snake_case : str ): if string == "True": return True elif string == "False": return False else: raise ValueError(F'could not parse string as bool {string}' ) parser.add_argument( '''--use_linear_projection''', help='''Override for use linear projection''', required=False, type=parse_bool ) parser.add_argument('''--cross_attention_dim''', help='''Override for cross attention_dim''', required=False, type=int) _UpperCAmelCase : Dict = parser.parse_args() _UpperCAmelCase : Optional[int] = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
107
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : str , A__ : str ) -> Optional[int]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : int , A__ : Optional[int] , A__ : Dict ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Any , A__ : List[str] , A__ : str ) -> Optional[int]: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : List[Any] , A__ : List[str] , A__ : Optional[int] ) -> List[str]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class snake_case__ ( _UpperCamelCase ): @require_beam def UpperCAmelCase__ ( self : str ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Dict = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[int] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : str ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam snake_case_ : Tuple = beam.io.parquetio.WriteToParquet snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[Any] = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: snake_case_ : int = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[Any] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Tuple = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[str] = NestedBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) snake_case_ : int = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
666
0
import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def _SCREAMING_SNAKE_CASE ( __snake_case ) -> Optional[int]: _UpperCAmelCase = np.inf def set_batch_size(__snake_case ) -> None: nonlocal batch_size if isinstance(__snake_case , __snake_case ): _UpperCAmelCase = min(__snake_case , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(__snake_case , __snake_case ): _UpperCAmelCase = min(__snake_case , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(__snake_case , __snake_case ) and feature.dtype == "binary": _UpperCAmelCase = min(__snake_case , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(__snake_case , __snake_case ) return None if batch_size is np.inf else batch_size class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any] , lowerCamelCase : NestedDataStructureLike[PathLike] , lowerCamelCase : Optional[NamedSplit] = None , lowerCamelCase : Optional[Features] = None , lowerCamelCase : str = None , lowerCamelCase : bool = False , lowerCamelCase : bool = False , lowerCamelCase : Optional[int] = None , **lowerCamelCase : Tuple , ) -> Optional[Any]: """simple docstring""" super().__init__( lowerCamelCase , split=lowerCamelCase , features=lowerCamelCase , cache_dir=lowerCamelCase , keep_in_memory=lowerCamelCase , streaming=lowerCamelCase , num_proc=lowerCamelCase , **lowerCamelCase , ) _UpperCAmelCase = path_or_paths if isinstance(lowerCamelCase , lowerCamelCase ) else {self.split: path_or_paths} _UpperCAmelCase = _PACKAGED_DATASETS_MODULES["""parquet"""][1] _UpperCAmelCase = Parquet( cache_dir=lowerCamelCase , data_files=lowerCamelCase , features=lowerCamelCase , hash=lowerCamelCase , **lowerCamelCase , ) def lowerCamelCase ( self : Optional[Any] ) -> List[str]: """simple docstring""" # Build iterable dataset if self.streaming: _UpperCAmelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None self.builder.download_and_prepare( download_config=lowerCamelCase , download_mode=lowerCamelCase , verification_mode=lowerCamelCase , base_path=lowerCamelCase , num_proc=self.num_proc , ) _UpperCAmelCase = self.builder.as_dataset( split=self.split , verification_mode=lowerCamelCase , in_memory=self.keep_in_memory ) return dataset class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : List[str] , lowerCamelCase : Dataset , lowerCamelCase : Union[PathLike, BinaryIO] , lowerCamelCase : Optional[int] = None , **lowerCamelCase : Any , ) -> Tuple: """simple docstring""" _UpperCAmelCase = dataset _UpperCAmelCase = path_or_buf _UpperCAmelCase = batch_size or get_writer_batch_size(dataset.features ) _UpperCAmelCase = parquet_writer_kwargs def lowerCamelCase ( self : List[Any] ) -> int: """simple docstring""" _UpperCAmelCase = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , """wb+""" ) as buffer: _UpperCAmelCase = self._write(file_obj=lowerCamelCase , batch_size=lowerCamelCase , **self.parquet_writer_kwargs ) else: _UpperCAmelCase = self._write(file_obj=self.path_or_buf , batch_size=lowerCamelCase , **self.parquet_writer_kwargs ) return written def lowerCamelCase ( self : Dict , lowerCamelCase : BinaryIO , lowerCamelCase : int , **lowerCamelCase : List[str] ) -> int: """simple docstring""" _UpperCAmelCase = 0 _UpperCAmelCase = parquet_writer_kwargs.pop("""path_or_buf""" , lowerCamelCase ) _UpperCAmelCase = self.dataset.features.arrow_schema _UpperCAmelCase = pq.ParquetWriter(lowerCamelCase , schema=lowerCamelCase , **lowerCamelCase ) for offset in logging.tqdm( range(0 , len(self.dataset ) , lowerCamelCase ) , unit="""ba""" , disable=not logging.is_progress_bar_enabled() , desc="""Creating parquet from Arrow format""" , ): _UpperCAmelCase = query_table( table=self.dataset._data , key=slice(lowerCamelCase , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(lowerCamelCase ) written += batch.nbytes writer.close() return written
108
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple ): class snake_case__ : def __init__( self : Any , A__ : List[str] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = metric_id class snake_case__ : _SCREAMING_SNAKE_CASE : List[str] = [MetricMock(_UpperCamelCase ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: int , lowerCAmelCase_: List[Any] , lowerCAmelCase_: Any , lowerCAmelCase_: List[str] ): if "tmp_path" in args: snake_case_ : List[Any] = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowerCAmelCase_ , match="https://huggingface.co/docs/evaluate" ): func(*lowerCAmelCase_ )
666
0
'''simple docstring''' from __future__ import annotations from math import gcd def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase = 2 , __UpperCAmelCase = 1 , __UpperCAmelCase = 3 , ) -> int | None: '''simple docstring''' if num < 2: raise ValueError("""The input value cannot be less than 2""" ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> int: return (pow(__UpperCAmelCase , 2 ) + step) % modulus for _ in range(__UpperCAmelCase ): # These track the position within the cycle detection logic. __SCREAMING_SNAKE_CASE = seed __SCREAMING_SNAKE_CASE = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. __SCREAMING_SNAKE_CASE = rand_fn(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) __SCREAMING_SNAKE_CASE = rand_fn(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) __SCREAMING_SNAKE_CASE = rand_fn(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. __SCREAMING_SNAKE_CASE = gcd(hare - tortoise , __UpperCAmelCase ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. __SCREAMING_SNAKE_CASE = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse a = argparse.ArgumentParser() parser.add_argument( "num", type=int, help="The value to find a divisor of", ) parser.add_argument( "--attempts", type=int, default=3, help="The number of attempts before giving up", ) a = parser.parse_args() a = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(F'''{args.num} is probably prime''') else: a = args.num // divisor print(F'''{args.num} = {divisor} * {quotient}''')
109
from __future__ import annotations import bisect def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Any = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: snake_case_ : Tuple = mid + 1 else: snake_case_ : Dict = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Optional[Any] = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: snake_case_ : Optional[Any] = mid + 1 else: snake_case_ : Tuple = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_left(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_right(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Dict = 0 snake_case_ : Tuple = len(lowerCAmelCase_ ) - 1 while left <= right: snake_case_ : int = left + (right - left) // 2 snake_case_ : Optional[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: snake_case_ : Optional[Any] = midpoint - 1 else: snake_case_ : Optional[int] = midpoint + 1 return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Optional[int] = bisect.bisect_left(lowerCAmelCase_ , lowerCAmelCase_ ) if index != len(lowerCAmelCase_ ) and sorted_collection[index] == item: return index return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int , lowerCAmelCase_: int ): if right < left: return None snake_case_ : List[Any] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , midpoint - 1 ) else: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , midpoint + 1 , lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = input("Enter numbers separated by comma:\n").strip() UpperCAmelCase = sorted(int(item) for item in user_input.split(",")) UpperCAmelCase = int(input("Enter a single number to be found in the list:\n")) UpperCAmelCase = binary_search(collection, target) if result is None: print(F"{target} was not found in {collection}.") else: print(F"{target} was found at position {result} in {collection}.")
666
0
import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def a_ ( SCREAMING_SNAKE_CASE__ : Dict ): '''simple docstring''' _lowerCamelCase : Optional[int] =VideoMAEConfig() set_architecture_configs(lowerCAmelCase_ , lowerCAmelCase_ ) if "finetuned" not in model_name: _lowerCamelCase : int =False if "finetuned" in model_name: _lowerCamelCase : int ="huggingface/label-files" if "kinetics" in model_name: _lowerCamelCase : List[str] =400 _lowerCamelCase : Dict ="kinetics400-id2label.json" elif "ssv2" in model_name: _lowerCamelCase : Dict =174 _lowerCamelCase : Optional[Any] ="something-something-v2-id2label.json" else: raise ValueError('Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.' ) _lowerCamelCase : Dict =json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type='dataset' ) , 'r' ) ) _lowerCamelCase : List[Any] ={int(lowerCAmelCase_ ): v for k, v in idalabel.items()} _lowerCamelCase : Any =idalabel _lowerCamelCase : List[Any] ={v: k for k, v in idalabel.items()} return config def a_ ( SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] ): '''simple docstring''' if "small" in model_name: _lowerCamelCase : Dict =384 _lowerCamelCase : List[str] =1_536 _lowerCamelCase : Dict =12 _lowerCamelCase : List[str] =16 _lowerCamelCase : str =12 _lowerCamelCase : int =3 _lowerCamelCase : List[str] =192 _lowerCamelCase : List[str] =768 elif "large" in model_name: _lowerCamelCase : Optional[int] =1_024 _lowerCamelCase : str =4_096 _lowerCamelCase : Any =24 _lowerCamelCase : str =16 _lowerCamelCase : str =12 _lowerCamelCase : Dict =8 _lowerCamelCase : Any =512 _lowerCamelCase : str =2_048 elif "huge" in model_name: _lowerCamelCase : Optional[int] =1_280 _lowerCamelCase : Tuple =5_120 _lowerCamelCase : Dict =32 _lowerCamelCase : Union[str, Any] =16 _lowerCamelCase : Union[str, Any] =12 _lowerCamelCase : Optional[Any] =8 _lowerCamelCase : Tuple =640 _lowerCamelCase : Any =2_560 elif "base" not in model_name: raise ValueError('Model name should include either \"small\", \"base\", \"large\", or \"huge\"' ) def a_ ( SCREAMING_SNAKE_CASE__ : Optional[int] ): '''simple docstring''' if "encoder." in name: _lowerCamelCase : Optional[int] =name.replace('encoder.' , '' ) if "cls_token" in name: _lowerCamelCase : Dict =name.replace('cls_token' , 'videomae.embeddings.cls_token' ) if "decoder_pos_embed" in name: _lowerCamelCase : str =name.replace('decoder_pos_embed' , 'decoder.decoder_pos_embed' ) if "pos_embed" in name and "decoder" not in name: _lowerCamelCase : List[str] =name.replace('pos_embed' , 'videomae.embeddings.position_embeddings' ) if "patch_embed.proj" in name: _lowerCamelCase : Dict =name.replace('patch_embed.proj' , 'videomae.embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowerCamelCase : Optional[Any] =name.replace('patch_embed.norm' , 'videomae.embeddings.norm' ) if "decoder.blocks" in name: _lowerCamelCase : Tuple =name.replace('decoder.blocks' , 'decoder.decoder_layers' ) if "blocks" in name: _lowerCamelCase : str =name.replace('blocks' , 'videomae.encoder.layer' ) if "attn.proj" in name: _lowerCamelCase : str =name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name and "bias" not in name: _lowerCamelCase : Optional[Any] =name.replace('attn' , 'attention.self' ) if "attn" in name: _lowerCamelCase : List[str] =name.replace('attn' , 'attention.attention' ) if "norm1" in name: _lowerCamelCase : Union[str, Any] =name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: _lowerCamelCase : int =name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: _lowerCamelCase : Optional[int] =name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _lowerCamelCase : Optional[int] =name.replace('mlp.fc2' , 'output.dense' ) if "decoder_embed" in name: _lowerCamelCase : List[str] =name.replace('decoder_embed' , 'decoder.decoder_embed' ) if "decoder_norm" in name: _lowerCamelCase : Optional[Any] =name.replace('decoder_norm' , 'decoder.decoder_norm' ) if "decoder_pred" in name: _lowerCamelCase : Any =name.replace('decoder_pred' , 'decoder.decoder_pred' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: _lowerCamelCase : Optional[int] =name.replace('norm.weight' , 'videomae.layernorm.weight' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: _lowerCamelCase : Any =name.replace('norm.bias' , 'videomae.layernorm.bias' ) if "head" in name and "decoder" not in name: _lowerCamelCase : str =name.replace('head' , 'classifier' ) return name def a_ ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase : str =orig_state_dict.pop(lowerCAmelCase_ ) if key.startswith('encoder.' ): _lowerCamelCase : Any =key.replace('encoder.' , '' ) if "qkv" in key: _lowerCamelCase : List[Any] =key.split('.' ) if key.startswith('decoder.blocks' ): _lowerCamelCase : Any =config.decoder_hidden_size _lowerCamelCase : List[Any] =int(key_split[2] ) _lowerCamelCase : Optional[Any] ="decoder.decoder_layers." if "weight" in key: _lowerCamelCase : Optional[Any] =val[:dim, :] _lowerCamelCase : Optional[int] =val[dim : dim * 2, :] _lowerCamelCase : Union[str, Any] =val[-dim:, :] else: _lowerCamelCase : List[Any] =config.hidden_size _lowerCamelCase : Union[str, Any] =int(key_split[1] ) _lowerCamelCase : Optional[Any] ="videomae.encoder.layer." if "weight" in key: _lowerCamelCase : Any =val[:dim, :] _lowerCamelCase : int =val[dim : dim * 2, :] _lowerCamelCase : Tuple =val[-dim:, :] else: _lowerCamelCase : Tuple =val return orig_state_dict def a_ ( ): '''simple docstring''' _lowerCamelCase : List[Any] =hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename='eating_spaghetti.npy' , repo_type='dataset' ) _lowerCamelCase : Any =np.load(lowerCAmelCase_ ) return list(lowerCAmelCase_ ) def a_ ( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Union[str, Any] ): '''simple docstring''' _lowerCamelCase : Optional[Any] =get_videomae_config(lowerCAmelCase_ ) if "finetuned" in model_name: _lowerCamelCase : Dict =VideoMAEForVideoClassification(lowerCAmelCase_ ) else: _lowerCamelCase : str =VideoMAEForPreTraining(lowerCAmelCase_ ) # download original checkpoint, hosted on Google Drive _lowerCamelCase : int ="pytorch_model.bin" gdown.cached_download(lowerCAmelCase_ , lowerCAmelCase_ , quiet=lowerCAmelCase_ ) _lowerCamelCase : str =torch.load(lowerCAmelCase_ , map_location='cpu' ) if "model" in files: _lowerCamelCase : str =files["model"] else: _lowerCamelCase : Any =files["module"] _lowerCamelCase : int =convert_state_dict(lowerCAmelCase_ , lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) model.eval() # verify model on basic input _lowerCamelCase : Any =VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) _lowerCamelCase : Optional[Any] =prepare_video() _lowerCamelCase : Union[str, Any] =image_processor(lowerCAmelCase_ , return_tensors='pt' ) if "finetuned" not in model_name: _lowerCamelCase : Optional[Any] =hf_hub_download(repo_id='hf-internal-testing/bool-masked-pos' , filename='bool_masked_pos.pt' ) _lowerCamelCase : Optional[int] =torch.load(lowerCAmelCase_ ) _lowerCamelCase : int =model(**lowerCAmelCase_ ) _lowerCamelCase : Tuple =outputs.logits _lowerCamelCase : str =[ "videomae-small-finetuned-kinetics", "videomae-small-finetuned-ssv2", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) "videomae-base-short", "videomae-base-short-finetuned-kinetics", "videomae-base", "videomae-base-finetuned-kinetics", "videomae-large", "videomae-large-finetuned-kinetics", "videomae-huge-finetuned-kinetics", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) "videomae-base-short-ssv2", "videomae-base-short-finetuned-ssv2", "videomae-base-ssv2", "videomae-base-finetuned-ssv2", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": _lowerCamelCase : Optional[Any] =torch.Size([1, 400] ) _lowerCamelCase : List[str] =torch.tensor([-0.92_91, -0.40_61, -0.93_07] ) elif model_name == "videomae-small-finetuned-ssv2": _lowerCamelCase : Tuple =torch.Size([1, 174] ) _lowerCamelCase : Optional[Any] =torch.tensor([0.26_71, -0.46_89, -0.82_35] ) elif model_name == "videomae-base": _lowerCamelCase : List[Any] =torch.Size([1, 1_408, 1_536] ) _lowerCamelCase : Optional[Any] =torch.tensor([[0.77_39, 0.79_68, 0.70_89], [0.67_01, 0.74_87, 0.62_09], [0.42_87, 0.51_58, 0.47_73]] ) elif model_name == "videomae-base-short": _lowerCamelCase : int =torch.Size([1, 1_408, 1_536] ) _lowerCamelCase : List[Any] =torch.tensor([[0.79_94, 0.96_12, 0.85_08], [0.74_01, 0.89_58, 0.83_02], [0.58_62, 0.74_68, 0.73_25]] ) # we verified the loss both for normalized and unnormalized targets for this one _lowerCamelCase : int =torch.tensor([0.51_42] ) if config.norm_pix_loss else torch.tensor([0.64_69] ) elif model_name == "videomae-large": _lowerCamelCase : str =torch.Size([1, 1_408, 1_536] ) _lowerCamelCase : Tuple =torch.tensor([[0.71_49, 0.79_97, 0.69_66], [0.67_68, 0.78_69, 0.69_48], [0.51_39, 0.62_21, 0.56_05]] ) elif model_name == "videomae-large-finetuned-kinetics": _lowerCamelCase : str =torch.Size([1, 400] ) _lowerCamelCase : int =torch.tensor([0.07_71, 0.00_11, -0.36_25] ) elif model_name == "videomae-huge-finetuned-kinetics": _lowerCamelCase : int =torch.Size([1, 400] ) _lowerCamelCase : Dict =torch.tensor([0.24_33, 0.16_32, -0.48_94] ) elif model_name == "videomae-base-short-finetuned-kinetics": _lowerCamelCase : Optional[int] =torch.Size([1, 400] ) _lowerCamelCase : Tuple =torch.tensor([0.65_88, 0.09_90, -0.24_93] ) elif model_name == "videomae-base-finetuned-kinetics": _lowerCamelCase : str =torch.Size([1, 400] ) _lowerCamelCase : Any =torch.tensor([0.36_69, -0.06_88, -0.24_21] ) elif model_name == "videomae-base-short-ssv2": _lowerCamelCase : int =torch.Size([1, 1_408, 1_536] ) _lowerCamelCase : str =torch.tensor([[0.47_12, 0.52_96, 0.57_86], [0.22_78, 0.27_29, 0.40_26], [0.03_52, 0.07_30, 0.25_06]] ) elif model_name == "videomae-base-short-finetuned-ssv2": _lowerCamelCase : List[str] =torch.Size([1, 174] ) _lowerCamelCase : List[str] =torch.tensor([-0.05_37, -0.15_39, -0.32_66] ) elif model_name == "videomae-base-ssv2": _lowerCamelCase : int =torch.Size([1, 1_408, 1_536] ) _lowerCamelCase : List[str] =torch.tensor([[0.81_31, 0.87_27, 0.85_46], [0.73_66, 0.93_77, 0.88_70], [0.59_35, 0.88_74, 0.85_64]] ) elif model_name == "videomae-base-finetuned-ssv2": _lowerCamelCase : Optional[Any] =torch.Size([1, 174] ) _lowerCamelCase : Any =torch.tensor([0.19_61, -0.83_37, -0.63_89] ) else: raise ValueError(F'''Model name not supported. Should be one of {model_names}''' ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , lowerCAmelCase_ , atol=1e-4 ) else: print('Logits:' , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , lowerCAmelCase_ , atol=1e-4 ) print('Logits ok!' ) # verify loss, if applicable if model_name == "videomae-base-short": _lowerCamelCase : Tuple =outputs.loss assert torch.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) print('Loss ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) if push_to_hub: print('Pushing to the hub...' ) model.push_to_hub(lowerCAmelCase_ , organization='nielsr' ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) lowerCamelCase = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
464
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , A__ : VQModel , A__ : UNetaDModel , A__ : DDIMScheduler ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , A__ : int = 1 , A__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A__ : float = 0.0 , A__ : int = 50 , A__ : Optional[str] = "pil" , A__ : bool = True , **A__ : Optional[Any] , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' snake_case_ : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) snake_case_ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler snake_case_ : Any = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature snake_case_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case_ : List[Any] = {} if accepts_eta: snake_case_ : int = eta for t in self.progress_bar(self.scheduler.timesteps ): snake_case_ : Union[str, Any] = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual snake_case_ : Dict = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case_ : Union[str, Any] = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE snake_case_ : int = self.vqvae.decode(A__ ).sample snake_case_ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ : Optional[int] = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
666
0
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class __magic_name__ : def __init__( self : Union[str, Any] , snake_case_ : Optional[Any] , snake_case_ : Optional[Any] , snake_case_ : Dict , snake_case_ : Union[str, Any] , snake_case_ : Union[str, Any] , snake_case_ : Optional[Any]=0.2 , snake_case_ : Any=0.2 ): __snake_case = bp_numa __snake_case = bp_numa __snake_case = bp_numa __snake_case = conva_get[:2] __snake_case = conva_get[2] __snake_case = size_pa __snake_case = rate_w __snake_case = rate_t __snake_case = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] __snake_case = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) __snake_case = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) __snake_case = -2 * np.random.rand(self.conva[1] ) + 1 __snake_case = -2 * np.random.rand(self.num_bpa ) + 1 __snake_case = -2 * np.random.rand(self.num_bpa ) + 1 def lowerCAmelCase ( self : Optional[Any] , snake_case_ : Optional[int] ): __snake_case = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(A__ , "wb" ) as f: pickle.dump(A__ , A__ ) print(F'''Model saved: {save_path}''' ) @classmethod def lowerCAmelCase ( cls : Optional[Any] , snake_case_ : Tuple ): with open(A__ , "rb" ) as f: __snake_case = pickle.load(A__ ) # noqa: S301 __snake_case = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) __snake_case = model_dic.get("size_pooling1" ) __snake_case = model_dic.get("num_bp1" ) __snake_case = model_dic.get("num_bp2" ) __snake_case = model_dic.get("num_bp3" ) __snake_case = model_dic.get("rate_weight" ) __snake_case = model_dic.get("rate_thre" ) # create model instance __snake_case = CNN(A__ , A__ , A__ , A__ , A__ , A__ , A__ ) # modify model parameter __snake_case = model_dic.get("w_conv1" ) __snake_case = model_dic.get("wkj" ) __snake_case = model_dic.get("vji" ) __snake_case = model_dic.get("thre_conv1" ) __snake_case = model_dic.get("thre_bp2" ) __snake_case = model_dic.get("thre_bp3" ) return conv_ins def lowerCAmelCase ( self : List[Any] , snake_case_ : Tuple ): return 1 / (1 + np.exp(-1 * x )) def lowerCAmelCase ( self : Any , snake_case_ : Tuple ): return round(A__ , 3 ) def lowerCAmelCase ( self : str , snake_case_ : Any , snake_case_ : Tuple , snake_case_ : Dict , snake_case_ : Union[str, Any] , snake_case_ : List[Any] ): __snake_case = convs[0] __snake_case = convs[1] __snake_case = np.shape(A__ )[0] # get the data slice of original image data, data_focus __snake_case = [] for i_focus in range(0 , size_data - size_conv + 1 , A__ ): for j_focus in range(0 , size_data - size_conv + 1 , A__ ): __snake_case = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(A__ ) # calculate the feature map of every single kernel, and saved as list of matrix __snake_case = [] __snake_case = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(A__ ): __snake_case = [] for i_focus in range(len(A__ ) ): __snake_case = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(A__ ) ) __snake_case = np.asmatrix(A__ ).reshape( A__ , A__ ) data_featuremap.append(A__ ) # expanding the data slice to One dimenssion __snake_case = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(A__ ) ) __snake_case = np.asarray(A__ ) return focus_list, data_featuremap def lowerCAmelCase ( self : str , snake_case_ : Dict , snake_case_ : Dict , snake_case_ : Union[str, Any]="average_pool" ): __snake_case = len(featuremaps[0] ) __snake_case = int(size_map / size_pooling ) __snake_case = [] for i_map in range(len(A__ ) ): __snake_case = featuremaps[i_map] __snake_case = [] for i_focus in range(0 , A__ , A__ ): for j_focus in range(0 , A__ , A__ ): __snake_case = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(A__ ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(A__ ) ) __snake_case = np.asmatrix(A__ ).reshape(A__ , A__ ) featuremap_pooled.append(A__ ) return featuremap_pooled def lowerCAmelCase ( self : List[str] , snake_case_ : Optional[int] ): __snake_case = [] for i in range(len(A__ ) ): __snake_case = np.shape(data[i] ) __snake_case = data[i].reshape(1 , shapes[0] * shapes[1] ) __snake_case = data_listed.getA().tolist()[0] data_expanded.extend(A__ ) __snake_case = np.asarray(A__ ) return data_expanded def lowerCAmelCase ( self : str , snake_case_ : List[str] ): __snake_case = np.asarray(A__ ) __snake_case = np.shape(A__ ) __snake_case = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def lowerCAmelCase ( self : Optional[int] , snake_case_ : Dict , snake_case_ : Union[str, Any] , snake_case_ : List[str] , snake_case_ : Union[str, Any] , snake_case_ : Optional[int] ): __snake_case = [] __snake_case = 0 for i_map in range(A__ ): __snake_case = np.ones((size_map, size_map) ) for i in range(0 , A__ , A__ ): for j in range(0 , A__ , A__ ): __snake_case = pd_pool[ i_pool ] __snake_case = i_pool + 1 __snake_case = np.multiply( A__ , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(A__ ) return pd_all def lowerCAmelCase ( self : Union[str, Any] , snake_case_ : Optional[int] , snake_case_ : Any , snake_case_ : str , snake_case_ : Optional[int] , snake_case_ : int , snake_case_ : int=bool ): print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(A__ )) ) print((" - - Shape: Teach_Data ", np.shape(A__ )) ) __snake_case = 0 __snake_case = [] __snake_case = 10000 while rp < n_repeat and mse >= error_accuracy: __snake_case = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(A__ ) ): # print('------------Learning Image: %d--------------'%p) __snake_case = np.asmatrix(datas_train[p] ) __snake_case = np.asarray(datas_teach[p] ) __snake_case = self.convolute( A__ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __snake_case = self.pooling(A__ , self.size_poolinga ) __snake_case = np.shape(A__ ) __snake_case = self._expand(A__ ) __snake_case = data_bp_input __snake_case = np.dot(A__ , self.vji.T ) - self.thre_bpa __snake_case = self.sig(A__ ) __snake_case = np.dot(A__ , self.wkj.T ) - self.thre_bpa __snake_case = self.sig(A__ ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- __snake_case = np.multiply( (data_teach - bp_outa) , np.multiply(A__ , (1 - bp_outa) ) ) __snake_case = np.multiply( np.dot(A__ , self.wkj ) , np.multiply(A__ , (1 - bp_outa) ) ) __snake_case = np.dot(A__ , self.vji ) __snake_case = pd_i_all / (self.size_poolinga * self.size_poolinga) __snake_case = pd_conva_pooled.T.getA().tolist() __snake_case = self._calculate_gradient_from_pool( A__ , A__ , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): __snake_case = self._expand_mat(pd_conva_all[k_conv] ) __snake_case = self.rate_weight * np.dot(A__ , A__ ) __snake_case = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) __snake_case = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer __snake_case = self.wkj + pd_k_all.T * bp_outa * self.rate_weight __snake_case = self.vji + pd_j_all.T * bp_outa * self.rate_weight __snake_case = self.thre_bpa - pd_k_all * self.rate_thre __snake_case = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image __snake_case = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) __snake_case = rp + 1 __snake_case = error_count / patterns all_mse.append(A__ ) def draw_error(): __snake_case = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(A__ , "+-" ) plt.plot(A__ , "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(A__ , alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def lowerCAmelCase ( self : Tuple , snake_case_ : Optional[int] ): __snake_case = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(A__ )) ) for p in range(len(A__ ) ): __snake_case = np.asmatrix(datas_test[p] ) __snake_case = self.convolute( A__ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __snake_case = self.pooling(A__ , self.size_poolinga ) __snake_case = self._expand(A__ ) __snake_case = data_bp_input __snake_case = bp_outa * self.vji.T - self.thre_bpa __snake_case = self.sig(A__ ) __snake_case = bp_outa * self.wkj.T - self.thre_bpa __snake_case = self.sig(A__ ) produce_out.extend(bp_outa.getA().tolist() ) __snake_case = [list(map(self.do_round , A__ ) ) for each in produce_out] return np.asarray(A__ ) def lowerCAmelCase ( self : Optional[Any] , snake_case_ : Dict ): __snake_case = np.asmatrix(A__ ) __snake_case = self.convolute( A__ , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) __snake_case = self.pooling(A__ , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
163
from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) snake_case_ : List[str] = precision snake_case_ : Union[str, Any] = ceil(precision / 1_4 ) snake_case_ : List[str] = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : str = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : str = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): snake_case_ : Tuple = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCAmelCase = 5_0 print(F"The first {n} digits of pi is: {pi(n)}")
666
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig class _a ( _UpperCamelCase ): a_ : Union[str, Any] = "bert-generation" def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[Any]=5_03_58 , SCREAMING_SNAKE_CASE__ : Any=10_24 , SCREAMING_SNAKE_CASE__ : Any=24 , SCREAMING_SNAKE_CASE__ : List[Any]=16 , SCREAMING_SNAKE_CASE__ : List[Any]=40_96 , SCREAMING_SNAKE_CASE__ : int="gelu" , SCREAMING_SNAKE_CASE__ : List[str]=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=0.1 , SCREAMING_SNAKE_CASE__ : str=5_12 , SCREAMING_SNAKE_CASE__ : int=0.02 , SCREAMING_SNAKE_CASE__ : Any=1e-12 , SCREAMING_SNAKE_CASE__ : Optional[Any]=0 , SCREAMING_SNAKE_CASE__ : List[str]=2 , SCREAMING_SNAKE_CASE__ : Optional[int]=1 , SCREAMING_SNAKE_CASE__ : str="absolute" , SCREAMING_SNAKE_CASE__ : Any=True , **SCREAMING_SNAKE_CASE__ : Optional[Any] , ): super().__init__(pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , **A__ ) lowerCamelCase__ = vocab_size lowerCamelCase__ = hidden_size lowerCamelCase__ = num_hidden_layers lowerCamelCase__ = num_attention_heads lowerCamelCase__ = hidden_act lowerCamelCase__ = intermediate_size lowerCamelCase__ = hidden_dropout_prob lowerCamelCase__ = attention_probs_dropout_prob lowerCamelCase__ = max_position_embeddings lowerCamelCase__ = initializer_range lowerCamelCase__ = layer_norm_eps lowerCamelCase__ = position_embedding_type lowerCamelCase__ = use_cache
510
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int = 1_0_0_0 ): snake_case_ ,snake_case_ : List[str] = 1, 1 snake_case_ : List[str] = 2 while True: snake_case_ : Tuple = 0 snake_case_ : Union[str, Any] = fa + fa snake_case_ ,snake_case_ : str = fa, f index += 1 for _ in str(lowerCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
666
0
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def _a ( self : Union[str, Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _a ( self : List[str] ): '''simple docstring''' A_ : Tuple = 1 A_ : List[Any] = 3 A_ : Union[str, Any] = (32, 32) A_ : List[Any] = floats_tensor((batch_size, num_channels) + sizes ,rng=random.Random(0 ) ).to(A__ ) return image @property def _a ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) A_ : Any = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") ,up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") ,cross_attention_dim=32 ,) return model @property def _a ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) A_ : Dict = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,) return model @property def _a ( self : int ): '''simple docstring''' torch.manual_seed(0 ) A_ : List[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(A__ ) @property def _a ( self : int ): '''simple docstring''' def extract(*_a : Dict ,**_a : Optional[int] ): class __lowerCAmelCase : '''simple docstring''' def __init__( self : List[str] ): '''simple docstring''' A_ : str = torch.ones([0] ) def _a ( self : Any ,_a : Dict ): '''simple docstring''' self.pixel_values.to(A__ ) return self return Out() return extract def _a ( self : Dict ): '''simple docstring''' A_ : int = "cpu" # ensure determinism for the device-dependent torch.Generator A_ : Dict = self.dummy_cond_unet A_ : int = DDIMScheduler( beta_start=0.00085 ,beta_end=0.012 ,beta_schedule="""scaled_linear""" ,clip_sample=A__ ,set_alpha_to_one=A__ ,) A_ : str = self.dummy_vae A_ : Any = self.dummy_text_encoder A_ : str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk A_ : Optional[Any] = StableDiffusionPipeline( unet=A__ ,scheduler=A__ ,vae=A__ ,text_encoder=A__ ,tokenizer=A__ ,safety_checker=A__ ,feature_extractor=self.dummy_extractor ,) A_ : Optional[Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : List[str] = "A painting of a squirrel eating a burger" A_ : int = torch.Generator(device=A__ ).manual_seed(0 ) A_ : Tuple = sd_pipe([prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ) A_ : Union[str, Any] = output.images A_ : Union[str, Any] = torch.Generator(device=A__ ).manual_seed(0 ) A_ : List[Any] = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ,return_dict=A__ ,)[0] A_ : Optional[Any] = image[0, -3:, -3:, -1] A_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A_ : List[str] = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _a ( self : List[Any] ): '''simple docstring''' A_ : int = "cpu" # ensure determinism for the device-dependent torch.Generator A_ : Optional[int] = self.dummy_cond_unet A_ : str = PNDMScheduler(skip_prk_steps=A__ ) A_ : List[Any] = self.dummy_vae A_ : Tuple = self.dummy_text_encoder A_ : List[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk A_ : Union[str, Any] = StableDiffusionPipeline( unet=A__ ,scheduler=A__ ,vae=A__ ,text_encoder=A__ ,tokenizer=A__ ,safety_checker=A__ ,feature_extractor=self.dummy_extractor ,) A_ : Optional[Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : Union[str, Any] = "A painting of a squirrel eating a burger" A_ : Optional[Any] = torch.Generator(device=A__ ).manual_seed(0 ) A_ : Tuple = sd_pipe([prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ) A_ : Optional[Any] = output.images A_ : int = torch.Generator(device=A__ ).manual_seed(0 ) A_ : Optional[Any] = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ,return_dict=A__ ,)[0] A_ : Dict = image[0, -3:, -3:, -1] A_ : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A_ : Any = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _a ( self : Tuple ): '''simple docstring''' A_ : Any = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" ,safety_checker=A__ ) assert isinstance(A__ ,A__ ) assert isinstance(pipe.scheduler ,A__ ) assert pipe.safety_checker is None A_ : Optional[Any] = pipe("""example prompt""" ,num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(A__ ) A_ : List[str] = StableDiffusionPipeline.from_pretrained(A__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None A_ : List[str] = pipe("""example prompt""" ,num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" ,"""This test requires a GPU""" ) def _a ( self : str ): '''simple docstring''' A_ : Union[str, Any] = self.dummy_cond_unet A_ : List[Any] = PNDMScheduler(skip_prk_steps=A__ ) A_ : str = self.dummy_vae A_ : Dict = self.dummy_text_encoder A_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 A_ : Dict = unet.half() A_ : List[str] = vae.half() A_ : str = bert.half() # make sure here that pndm scheduler skips prk A_ : str = StableDiffusionPipeline( unet=A__ ,scheduler=A__ ,vae=A__ ,text_encoder=A__ ,tokenizer=A__ ,safety_checker=A__ ,feature_extractor=self.dummy_extractor ,) A_ : List[str] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : List[str] = "A painting of a squirrel eating a burger" A_ : Optional[Any] = sd_pipe([prompt] ,num_inference_steps=2 ,output_type="""np""" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def _a ( self : Optional[int] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _a ( self : List[str] ): '''simple docstring''' A_ : Dict = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ,safety_checker=A__ ) A_ : Tuple = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) A_ : Dict = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : int = ( "portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle" " coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with" " anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and" " children from bahnhof zoo, detailed " ) A_ : Tuple = 4003660346 A_ : Optional[int] = 7 # without safety guidance (sld_guidance_scale = 0) A_ : Optional[Any] = torch.manual_seed(A__ ) A_ : Optional[Any] = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) A_ : Optional[int] = output.images A_ : Optional[int] = image[0, -3:, -3:, -1] A_ : Dict = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) A_ : int = torch.manual_seed(A__ ) A_ : Any = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) A_ : str = output.images A_ : Optional[int] = image[0, -3:, -3:, -1] A_ : Tuple = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _a ( self : Optional[int] ): '''simple docstring''' A_ : Optional[Any] = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ,safety_checker=A__ ) A_ : Optional[Any] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) A_ : Optional[int] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : int = "padme amidala taking a bath artwork, safe for work, no nudity" A_ : List[str] = 2734971755 A_ : str = 7 A_ : Optional[int] = torch.manual_seed(A__ ) A_ : Dict = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) A_ : List[str] = output.images A_ : Optional[int] = image[0, -3:, -3:, -1] A_ : int = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 A_ : Optional[Any] = torch.manual_seed(A__ ) A_ : Union[str, Any] = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) A_ : List[Any] = output.images A_ : Dict = image[0, -3:, -3:, -1] A_ : str = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _a ( self : int ): '''simple docstring''' A_ : int = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) A_ : Optional[Any] = sd_pipe.to(A__ ) sd_pipe.set_progress_bar_config(disable=A__ ) A_ : List[Any] = ( "the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c." " leyendecker" ) A_ : List[Any] = 1044355234 A_ : Tuple = 12 A_ : List[str] = torch.manual_seed(A__ ) A_ : Optional[Any] = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) A_ : Optional[int] = output.images A_ : int = image[0, -3:, -3:, -1] A_ : str = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 A_ : Dict = torch.manual_seed(A__ ) A_ : int = sd_pipe( [prompt] ,generator=A__ ,guidance_scale=A__ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) A_ : List[Any] = output.images A_ : Union[str, Any] = image[0, -3:, -3:, -1] A_ : int = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
665
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int | float] , lowerCAmelCase_: int , lowerCAmelCase_: int ): if len(lowerCAmelCase_ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(lowerCAmelCase_ ) or left < -len(lowerCAmelCase_ ) or right >= len(lowerCAmelCase_ ) or right < -len(lowerCAmelCase_ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case_ : List[Any] = (left + right) >> 1 # the middle snake_case_ : Dict = find_max(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # find max in range[left, mid] snake_case_ : int = find_max(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
666
0
'''simple docstring''' import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging __A =logging.get_logger(__name__) logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): if "xprophetnet" in prophetnet_checkpoint_path: UpperCAmelCase__ : int = XLMProphetNetForConditionalGenerationOld.from_pretrained(lowerCAmelCase_ ) UpperCAmelCase__ : Dict = XLMProphetNetForConditionalGeneration.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ ) else: UpperCAmelCase__ : Optional[int] = ProphetNetForConditionalGenerationOld.from_pretrained(lowerCAmelCase_ ) UpperCAmelCase__ : Union[str, Any] = ProphetNetForConditionalGeneration.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ ) UpperCAmelCase__ : Tuple = ["key_proj", "value_proj", "query_proj"] UpperCAmelCase__ : str = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: UpperCAmelCase__ : Any = key.split(""".""" ) if attributes[0] == "lm_head": UpperCAmelCase__ : List[Any] = prophet UpperCAmelCase__ : int = prophet_old else: UpperCAmelCase__ : Dict = prophet.prophetnet UpperCAmelCase__ : List[Any] = prophet_old.model UpperCAmelCase__ : str = False for attribute in attributes: if attribute in mapping: UpperCAmelCase__ : Dict = mapping[attribute] if not hasattr(lowerCAmelCase_ , lowerCAmelCase_ ) and len(lowerCAmelCase_ ) > 0: UpperCAmelCase__ : Any = attribute elif hasattr(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase__ : Tuple = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" UpperCAmelCase__ : List[str] = old_model.weight logger.info(f'''{attribute} is initialized.''' ) UpperCAmelCase__ : Optional[int] = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" UpperCAmelCase__ : str = old_model.bias logger.info(f'''{attribute} is initialized''' ) UpperCAmelCase__ : Optional[int] = True break elif attribute in special_keys and hasattr(lowerCAmelCase_ , """in_proj_weight""" ): UpperCAmelCase__ : Any = old_model.in_proj_weight.shape[0] // 3 UpperCAmelCase__ : Union[str, Any] = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": UpperCAmelCase__ : Dict = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) UpperCAmelCase__ : Tuple = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": UpperCAmelCase__ : Optional[int] = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) UpperCAmelCase__ : List[Any] = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": UpperCAmelCase__ : int = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) UpperCAmelCase__ : Optional[Any] = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) UpperCAmelCase__ : Any = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_1_2, "We want 512 position_embeddings." UpperCAmelCase__ : Any = nn.Parameter(old_model.embed_positions.weight[:5_1_2, :] ) UpperCAmelCase__ : Tuple = True break if attribute.isdigit(): UpperCAmelCase__ : int = model[int(lowerCAmelCase_ )] UpperCAmelCase__ : List[Any] = old_model[int(lowerCAmelCase_ )] else: UpperCAmelCase__ : int = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) if old_attribute == "": UpperCAmelCase__ : Optional[int] = old_model else: if not hasattr(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError(f'''{old_model} does not have {old_attribute}''' ) UpperCAmelCase__ : List[str] = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) if not is_key_init: raise ValueError(f'''{key} was not correctly initialized!''' ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) prophet.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": __A =argparse.ArgumentParser() # Required parameters parser.add_argument( '--prophetnet_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __A =parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
407
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } UpperCAmelCase = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE : List[str] = RobertaTokenizer def __init__( self : Optional[int] , A__ : List[Any]=None , A__ : Optional[int]=None , A__ : List[str]=None , A__ : Dict="replace" , A__ : List[str]="<s>" , A__ : Optional[Any]="</s>" , A__ : List[str]="</s>" , A__ : List[Any]="<s>" , A__ : int="<unk>" , A__ : int="<pad>" , A__ : List[Any]="<mask>" , A__ : Any=False , A__ : Optional[int]=True , **A__ : Union[str, Any] , ) -> int: '''simple docstring''' super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : List[Any] = getattr(A__ , pre_tok_state.pop("type" ) ) snake_case_ : Any = add_prefix_space snake_case_ : List[Any] = pre_tok_class(**A__ ) snake_case_ : Optional[int] = add_prefix_space snake_case_ : List[str] = "post_processor" snake_case_ : Tuple = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case_ : str = tuple(state["sep"] ) if "cls" in state: snake_case_ : Tuple = tuple(state["cls"] ) snake_case_ : Tuple = False if state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : Optional[Any] = add_prefix_space snake_case_ : str = True if state.get("trim_offsets" , A__ ) != trim_offsets: snake_case_ : Optional[int] = trim_offsets snake_case_ : List[Any] = True if changes_to_apply: snake_case_ : int = getattr(A__ , state.pop("type" ) ) snake_case_ : List[Any] = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCAmelCase__ ( self : Optional[Any] ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCAmelCase__ ( self : Tuple , A__ : Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case_ : Any = value def UpperCAmelCase__ ( self : int , *A__ : Optional[Any] , **A__ : int ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[Any] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , *A__ : Any , **A__ : List[Any] ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[int] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCAmelCase__ ( self : int , A__ : List[str] , A__ : Union[str, Any]=None ) -> Any: '''simple docstring''' snake_case_ : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Dict , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : str = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
666
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = { "configuration_roformer": ["ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoFormerConfig", "RoFormerOnnxConfig"], "tokenization_roformer": ["RoFormerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Tuple = ["RoFormerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Tuple = [ "ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "RoFormerForCausalLM", "RoFormerForMaskedLM", "RoFormerForMultipleChoice", "RoFormerForQuestionAnswering", "RoFormerForSequenceClassification", "RoFormerForTokenClassification", "RoFormerLayer", "RoFormerModel", "RoFormerPreTrainedModel", "load_tf_weights_in_roformer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[Any] = [ "TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRoFormerForCausalLM", "TFRoFormerForMaskedLM", "TFRoFormerForMultipleChoice", "TFRoFormerForQuestionAnswering", "TFRoFormerForSequenceClassification", "TFRoFormerForTokenClassification", "TFRoFormerLayer", "TFRoFormerModel", "TFRoFormerPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : str = [ "FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxRoFormerForMaskedLM", "FlaxRoFormerForMultipleChoice", "FlaxRoFormerForQuestionAnswering", "FlaxRoFormerForSequenceClassification", "FlaxRoFormerForTokenClassification", "FlaxRoFormerModel", "FlaxRoFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
298
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) snake_case_ : Any = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } snake_case_ : List[str] = model(A__ )["last_hidden_state"] snake_case_ : str = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , A__ ) # compare the actual values for a slice. snake_case_ : List[str] = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
666
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType lowerCamelCase : str = logging.get_logger(__name__) lowerCamelCase : Dict = { '''openai/whisper-base''': '''https://huggingface.co/openai/whisper-base/resolve/main/config.json''', } # fmt: off lowerCamelCase : Tuple = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_57, 3_66, 4_38, 5_32, 6_85, 7_05, 7_96, 9_30, 10_58, 12_20, 12_67, 12_79, 13_03, 13_43, 13_77, 13_91, 16_35, 17_82, 18_75, 21_62, 23_61, 24_88, 34_67, 40_08, 42_11, 46_00, 48_08, 52_99, 58_55, 63_29, 72_03, 96_09, 99_59, 1_05_63, 1_07_86, 1_14_20, 1_17_09, 1_19_07, 1_31_63, 1_36_97, 1_37_00, 1_48_08, 1_53_06, 1_64_10, 1_67_91, 1_79_92, 1_92_03, 1_95_10, 2_07_24, 2_23_05, 2_29_35, 2_70_07, 3_01_09, 3_04_20, 3_34_09, 3_49_49, 4_02_83, 4_04_93, 4_05_49, 4_72_82, 4_91_46, 5_02_57, 5_03_59, 5_03_60, 5_03_61 ] lowerCamelCase : List[str] = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_59, 5_03, 5_22, 5_42, 8_73, 8_93, 9_02, 9_18, 9_22, 9_31, 13_50, 18_53, 19_82, 24_60, 26_27, 32_46, 32_53, 32_68, 35_36, 38_46, 39_61, 41_83, 46_67, 65_85, 66_47, 72_73, 90_61, 93_83, 1_04_28, 1_09_29, 1_19_38, 1_20_33, 1_23_31, 1_25_62, 1_37_93, 1_41_57, 1_46_35, 1_52_65, 1_56_18, 1_65_53, 1_66_04, 1_83_62, 1_89_56, 2_00_75, 2_16_75, 2_25_20, 2_61_30, 2_61_61, 2_64_35, 2_82_79, 2_94_64, 3_16_50, 3_23_02, 3_24_70, 3_68_65, 4_28_63, 4_74_25, 4_98_70, 5_02_54, 5_02_58, 5_03_60, 5_03_61, 5_03_62 ] class lowerCAmelCase ( _UpperCamelCase ): '''simple docstring''' _A : Optional[Any] = "whisper" _A : Union[str, Any] = ["past_key_values"] _A : List[Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : int , __a : Any=51865 , __a : List[Any]=80 , __a : Any=6 , __a : int=4 , __a : Tuple=6 , __a : Optional[int]=4 , __a : int=1536 , __a : int=1536 , __a : int=0.0 , __a : List[str]=0.0 , __a : str=50257 , __a : List[Any]=True , __a : Optional[Any]=True , __a : Any="gelu" , __a : Tuple=256 , __a : List[Any]=0.0 , __a : Union[str, Any]=0.0 , __a : Any=0.0 , __a : List[Any]=0.02 , __a : Dict=False , __a : str=1500 , __a : List[str]=448 , __a : str=50256 , __a : str=50256 , __a : int=50256 , __a : Any=None , __a : Tuple=[220, 50256] , __a : Dict=False , __a : Union[str, Any]=256 , __a : Optional[int]=False , __a : Tuple=0.05 , __a : Tuple=10 , __a : Union[str, Any]=2 , __a : List[str]=0.0 , __a : List[str]=10 , __a : Optional[int]=0 , __a : Any=7 , **__a : Tuple , ) -> Any: """simple docstring""" __lowercase : Optional[Any] = vocab_size __lowercase : Optional[Any] = num_mel_bins __lowercase : Dict = d_model __lowercase : str = encoder_layers __lowercase : str = encoder_attention_heads __lowercase : Dict = decoder_layers __lowercase : List[Any] = decoder_attention_heads __lowercase : int = decoder_ffn_dim __lowercase : Optional[int] = encoder_ffn_dim __lowercase : str = dropout __lowercase : str = attention_dropout __lowercase : Dict = activation_dropout __lowercase : Union[str, Any] = activation_function __lowercase : Dict = init_std __lowercase : Optional[int] = encoder_layerdrop __lowercase : int = decoder_layerdrop __lowercase : Tuple = use_cache __lowercase : Any = encoder_layers __lowercase : Dict = scale_embedding # scale factor will be sqrt(d_model) if True __lowercase : Optional[Any] = max_source_positions __lowercase : Dict = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowercase : int = classifier_proj_size __lowercase : Any = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase : str = apply_spec_augment __lowercase : List[str] = mask_time_prob __lowercase : Tuple = mask_time_length __lowercase : Union[str, Any] = mask_time_min_masks __lowercase : int = mask_feature_prob __lowercase : Any = mask_feature_length __lowercase : str = mask_feature_min_masks __lowercase : Tuple = median_filter_width super().__init__( pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , is_encoder_decoder=A__ , decoder_start_token_id=A__ , suppress_tokens=A__ , begin_suppress_tokens=A__ , **A__ , ) class lowerCAmelCase ( _UpperCamelCase ): '''simple docstring''' @property def lowerCAmelCase ( self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowercase : Optional[Any] = OrderedDict( [ ("""input_features""", {0: """batch""", 1: """feature_size""", 2: """encoder_sequence"""}), ] ) if self.use_past: __lowercase : List[str] = {0: "batch"} else: __lowercase : str = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(A__ , direction="""inputs""" ) return common_inputs def lowerCAmelCase ( self : Dict , __a : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , __a : int = -1 , __a : int = -1 , __a : bool = False , __a : Optional["TensorType"] = None , __a : int = 22050 , __a : float = 5.0 , __a : int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowercase : Any = OrderedDict() __lowercase : Optional[int] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=A__ , framework=A__ , sampling_rate=A__ , time_duration=A__ , frequency=A__ , ) __lowercase : Optional[Any] = encoder_inputs["input_features"].shape[2] __lowercase : Any = encoder_sequence_length // 2 if self.use_past else seq_length __lowercase : Any = super().generate_dummy_inputs( preprocessor.tokenizer , A__ , A__ , A__ , A__ ) __lowercase : Any = encoder_inputs.pop("""input_features""" ) __lowercase : str = decoder_inputs.pop("""decoder_input_ids""" ) if "past_key_values" in decoder_inputs: __lowercase : Any = decoder_inputs.pop("""past_key_values""" ) return dummy_inputs @property def lowerCAmelCase ( self : List[str] ) -> float: """simple docstring""" return 1E-3
149
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = ["pixel_values"] def __init__( self : Union[str, Any] , A__ : bool = True , A__ : Dict[str, int] = None , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = True , **A__ : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : str = size if size is not None else {"shortest_edge": 2_24} snake_case_ : Union[str, Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : str = size snake_case_ : Optional[Any] = resample snake_case_ : Any = do_center_crop snake_case_ : Any = crop_size snake_case_ : str = do_rescale snake_case_ : Optional[Any] = rescale_factor snake_case_ : int = do_normalize snake_case_ : Optional[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ : List[str] = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ : int = do_convert_rgb def UpperCAmelCase__ ( self : Optional[int] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : str = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : np.ndarray , A__ : Union[int, float] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> str: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Any , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Any , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[Any] , A__ : ImageInput , A__ : bool = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : int = None , A__ : bool = None , A__ : float = None , A__ : bool = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A__ : Optional[Any] , ) -> PIL.Image.Image: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Union[str, Any] = size if size is not None else self.size snake_case_ : Any = get_size_dict(A__ , param_name="size" , default_to_square=A__ ) snake_case_ : Optional[int] = resample if resample is not None else self.resample snake_case_ : int = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" , default_to_square=A__ ) snake_case_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : List[Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ : List[Any] = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ : Dict = [convert_to_rgb(A__ ) for image in images] # All transformations expect numpy arrays. snake_case_ : Dict = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : Tuple = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : int = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : List[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Tuple = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ )
666
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ : Optional[int] = logging.get_logger(__name__) def __SCREAMING_SNAKE_CASE ( a__ : Union[str, Any] ) -> str: __A : Any = SwinConfig( embed_dim=192 ,depths=(2, 2, 18, 2) ,num_heads=(6, 12, 24, 48) ,window_size=12 ,out_features=["""stage2""", """stage3""", """stage4"""] ,) __A : Tuple = DetaConfig( backbone_config=lowerCAmelCase_ ,num_queries=900 ,encoder_ffn_dim=2048 ,decoder_ffn_dim=2048 ,num_feature_levels=5 ,assign_first_stage=lowerCAmelCase_ ,with_box_refine=lowerCAmelCase_ ,two_stage=lowerCAmelCase_ ,) # set labels __A : int = "huggingface/label-files" if "o365" in model_name: __A : List[Any] = 366 __A : int = "object365-id2label.json" else: __A : Optional[Any] = 91 __A : Any = "coco-detection-id2label.json" __A : str = num_labels __A : Union[str, Any] = json.load(open(cached_download(hf_hub_url(lowerCAmelCase_ ,lowerCAmelCase_ ,repo_type="""dataset""" ) ) ,"""r""" ) ) __A : int = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} __A : str = idalabel __A : Any = {v: k for k, v in idalabel.items()} return config def __SCREAMING_SNAKE_CASE ( a__ : int ) -> Optional[Any]: __A : Dict = [] # stem # fmt: off rename_keys.append(("""backbone.0.body.patch_embed.proj.weight""", """model.backbone.model.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""backbone.0.body.patch_embed.proj.bias""", """model.backbone.model.embeddings.patch_embeddings.projection.bias""") ) rename_keys.append(("""backbone.0.body.patch_embed.norm.weight""", """model.backbone.model.embeddings.norm.weight""") ) rename_keys.append(("""backbone.0.body.patch_embed.norm.bias""", """model.backbone.model.embeddings.norm.bias""") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.norm1.weight""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.norm1.bias""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.norm2.weight""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.norm2.bias""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias""", f"""model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((f"""backbone.0.body.layers.{i}.downsample.reduction.weight""", f"""model.backbone.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.downsample.norm.weight""", f"""model.backbone.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((f"""backbone.0.body.layers.{i}.downsample.norm.bias""", f"""model.backbone.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append(("""backbone.0.body.norm1.weight""", """model.backbone.model.hidden_states_norms.stage2.weight""") ) rename_keys.append(("""backbone.0.body.norm1.bias""", """model.backbone.model.hidden_states_norms.stage2.bias""") ) rename_keys.append(("""backbone.0.body.norm2.weight""", """model.backbone.model.hidden_states_norms.stage3.weight""") ) rename_keys.append(("""backbone.0.body.norm2.bias""", """model.backbone.model.hidden_states_norms.stage3.bias""") ) rename_keys.append(("""backbone.0.body.norm3.weight""", """model.backbone.model.hidden_states_norms.stage4.weight""") ) rename_keys.append(("""backbone.0.body.norm3.bias""", """model.backbone.model.hidden_states_norms.stage4.bias""") ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight""", f"""model.encoder.layers.{i}.self_attn.sampling_offsets.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias""", f"""model.encoder.layers.{i}.self_attn.sampling_offsets.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.attention_weights.weight""", f"""model.encoder.layers.{i}.self_attn.attention_weights.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.attention_weights.bias""", f"""model.encoder.layers.{i}.self_attn.attention_weights.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.value_proj.weight""", f"""model.encoder.layers.{i}.self_attn.value_proj.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.value_proj.bias""", f"""model.encoder.layers.{i}.self_attn.value_proj.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.output_proj.weight""", f"""model.encoder.layers.{i}.self_attn.output_proj.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.self_attn.output_proj.bias""", f"""model.encoder.layers.{i}.self_attn.output_proj.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.norm1.weight""", f"""model.encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.norm1.bias""", f"""model.encoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.linear1.weight""", f"""model.encoder.layers.{i}.fc1.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.linear1.bias""", f"""model.encoder.layers.{i}.fc1.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.linear2.weight""", f"""model.encoder.layers.{i}.fc2.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.linear2.bias""", f"""model.encoder.layers.{i}.fc2.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.norm2.weight""", f"""model.encoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.norm2.bias""", f"""model.encoder.layers.{i}.final_layer_norm.bias""") ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight""", f"""model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias""", f"""model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.attention_weights.weight""", f"""model.decoder.layers.{i}.encoder_attn.attention_weights.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.attention_weights.bias""", f"""model.decoder.layers.{i}.encoder_attn.attention_weights.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.value_proj.weight""", f"""model.decoder.layers.{i}.encoder_attn.value_proj.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.value_proj.bias""", f"""model.decoder.layers.{i}.encoder_attn.value_proj.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.output_proj.weight""", f"""model.decoder.layers.{i}.encoder_attn.output_proj.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.cross_attn.output_proj.bias""", f"""model.decoder.layers.{i}.encoder_attn.output_proj.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm1.weight""", f"""model.decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm1.bias""", f"""model.decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", f"""model.decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", f"""model.decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm2.weight""", f"""model.decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm2.bias""", f"""model.decoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.linear1.weight""", f"""model.decoder.layers.{i}.fc1.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.linear1.bias""", f"""model.decoder.layers.{i}.fc1.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.linear2.weight""", f"""model.decoder.layers.{i}.fc2.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.linear2.bias""", f"""model.decoder.layers.{i}.fc2.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm3.weight""", f"""model.decoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm3.bias""", f"""model.decoder.layers.{i}.final_layer_norm.bias""") ) # fmt: on return rename_keys def __SCREAMING_SNAKE_CASE ( a__ : str ,a__ : Optional[Any] ,a__ : List[str] ) -> int: __A : Dict = dct.pop(lowerCAmelCase_ ) __A : Dict = val def __SCREAMING_SNAKE_CASE ( a__ : Optional[Any] ,a__ : List[Any] ) -> str: __A : int = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __A : str = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __A : Optional[Any] = state_dict.pop(f"""backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight""" ) __A : str = state_dict.pop(f"""backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __A : str = in_proj_weight[:dim, :] __A : List[str] = in_proj_bias[: dim] __A : Optional[int] = in_proj_weight[ dim : dim * 2, : ] __A : Any = in_proj_bias[ dim : dim * 2 ] __A : Optional[Any] = in_proj_weight[ -dim :, : ] __A : Optional[Any] = in_proj_bias[-dim :] # fmt: on def __SCREAMING_SNAKE_CASE ( a__ : Dict ,a__ : Union[str, Any] ) -> Union[str, Any]: # transformer decoder self-attention layers __A : str = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention __A : List[Any] = state_dict.pop(f"""transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) __A : str = state_dict.pop(f"""transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __A : Optional[Any] = in_proj_weight[:hidden_size, :] __A : Optional[Any] = in_proj_bias[:hidden_size] __A : Optional[Any] = in_proj_weight[ hidden_size : hidden_size * 2, : ] __A : Union[str, Any] = in_proj_bias[hidden_size : hidden_size * 2] __A : Any = in_proj_weight[-hidden_size:, :] __A : Optional[Any] = in_proj_bias[-hidden_size:] def __SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: __A : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" __A : Tuple = Image.open(requests.get(lowerCAmelCase_ ,stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def __SCREAMING_SNAKE_CASE ( a__ : Union[str, Any] ,a__ : int ,a__ : Optional[Any] ) -> int: __A : Any = get_deta_config(lowerCAmelCase_ ) # load original state dict if model_name == "deta-swin-large": __A : int = hf_hub_download(repo_id="""nielsr/deta-checkpoints""" ,filename="""adet_swin_ft.pth""" ) elif model_name == "deta-swin-large-o365": __A : Dict = hf_hub_download(repo_id="""jozhang97/deta-swin-l-o365""" ,filename="""deta_swin_pt_o365.pth""" ) else: raise ValueError(f"""Model name {model_name} not supported""" ) __A : int = torch.load(lowerCAmelCase_ ,map_location="""cpu""" )["model"] # original state dict for name, param in state_dict.items(): print(lowerCAmelCase_ ,param.shape ) # rename keys __A : str = create_rename_keys(lowerCAmelCase_ ) for src, dest in rename_keys: rename_key(lowerCAmelCase_ ,lowerCAmelCase_ ,lowerCAmelCase_ ) read_in_swin_q_k_v(lowerCAmelCase_ ,config.backbone_config ) read_in_decoder_q_k_v(lowerCAmelCase_ ,lowerCAmelCase_ ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: __A : Dict = state_dict.pop(lowerCAmelCase_ ) __A : Optional[int] = val if "input_proj" in key: __A : Optional[Any] = state_dict.pop(lowerCAmelCase_ ) __A : Dict = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: __A : Union[str, Any] = state_dict.pop(lowerCAmelCase_ ) __A : str = val # finally, create HuggingFace model and load state dict __A : int = DetaForObjectDetection(lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) model.eval() __A : Any = "cuda" if torch.cuda.is_available() else "cpu" model.to(lowerCAmelCase_ ) # load image processor __A : List[str] = DetaImageProcessor(format="""coco_detection""" ) # verify our conversion on image __A : Optional[Any] = prepare_img() __A : int = processor(images=lowerCAmelCase_ ,return_tensors="""pt""" ) __A : List[Any] = encoding["pixel_values"] __A : List[Any] = model(pixel_values.to(lowerCAmelCase_ ) ) # verify logits print("""Logits:""" ,outputs.logits[0, :3, :3] ) print("""Boxes:""" ,outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": __A : Any = torch.tensor( [[-7.6_308, -2.8_485, -5.3_737], [-7.2_037, -4.5_505, -4.8_027], [-7.2_943, -4.2_611, -4.6_617]] ) __A : Any = torch.tensor([[0.4_987, 0.4_969, 0.9_999], [0.2_549, 0.5_498, 0.4_805], [0.5_498, 0.2_757, 0.0_569]] ) elif model_name == "deta-swin-large-o365": __A : List[str] = torch.tensor( [[-8.0_122, -3.5_720, -4.9_717], [-8.1_547, -3.6_886, -4.6_389], [-7.6_610, -3.6_194, -5.0_134]] ) __A : str = torch.tensor([[0.2_523, 0.5_549, 0.4_881], [0.7_715, 0.4_149, 0.4_601], [0.5_503, 0.2_753, 0.0_575]] ) assert torch.allclose(outputs.logits[0, :3, :3] ,expected_logits.to(lowerCAmelCase_ ) ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] ,expected_boxes.to(lowerCAmelCase_ ) ,atol=1E-4 ) print("""Everything ok!""" ) if pytorch_dump_folder_path: # Save model and processor logger.info(f"""Saving PyTorch model and processor to {pytorch_dump_folder_path}...""" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) # Push to hub if push_to_hub: print("""Pushing model and processor to hub...""" ) model.push_to_hub(f"""jozhang97/{model_name}""" ) processor.push_to_hub(f"""jozhang97/{model_name}""" ) if __name__ == "__main__": UpperCAmelCase_ : List[str] = argparse.ArgumentParser() parser.add_argument( '''--model_name''', type=str, default='''deta-swin-large''', choices=['''deta-swin-large''', '''deta-swin-large-o365'''], help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) UpperCAmelCase_ : Optional[int] = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
17
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): snake_case_ ,snake_case_ : Dict = position snake_case_ : int = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] snake_case_ : Union[str, Any] = [] for position in positions: snake_case_ ,snake_case_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(lowerCAmelCase_ ) return permissible_positions def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] , lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): if is_complete(lowerCAmelCase_ ): return True for position in get_valid_pos(lowerCAmelCase_ , len(lowerCAmelCase_ ) ): snake_case_ ,snake_case_ : Dict = position if board[y][x] == 0: snake_case_ : List[str] = curr + 1 if open_knight_tour_helper(lowerCAmelCase_ , lowerCAmelCase_ , curr + 1 ): return True snake_case_ : Dict = 0 return False def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): snake_case_ : Optional[Any] = 1 if open_knight_tour_helper(lowerCAmelCase_ , (i, j) , 1 ): return board snake_case_ : Dict = 0 snake_case_ : str = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
666
0
"""simple docstring""" import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( _UpperCamelCase , unittest.TestCase ): lowerCamelCase : Dict =GPTSanJapaneseTokenizer lowerCamelCase : Optional[int] =False lowerCamelCase : Optional[int] ={"do_clean_text": False, "add_prefix_space": False} def __a ( self ) -> Optional[Any]: super().setUp() # fmt: off a : Tuple = ["こん", "こんに", "にちは", "ばんは", "世界,㔺界", "、", "。", "<BR>", "<SP>", "<TAB>", "<URL>", "<EMAIL>", "<TEL>", "<DATE>", "<PRICE>", "<BLOCK>", "<KIGOU>", "<U2000U2BFF>", "<|emoji1|>", "<unk>", "<|bagoftoken|>", "<|endoftext|>"] # fmt: on a : List[str] = {"emoji": {"\ud83d\ude00": "<|emoji1|>"}, "emoji_inv": {"<|emoji1|>": "\ud83d\ude00"}} # 😀 a : Dict = {"unk_token": "<unk>"} a : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) a : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["emoji_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) with open(self.emoji_file , "w" ) as emoji_writer: emoji_writer.write(json.dumps(A__ ) ) def __a ( self , **lowerCAmelCase__ ) -> List[str]: kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **A__ ) def __a ( self , lowerCAmelCase__ ) -> str: a : List[str] = "こんにちは、世界。 \nこんばんは、㔺界。😀" a : List[Any] = "こんにちは、世界。 \nこんばんは、世界。😀" return input_text, output_text def __a ( self , lowerCAmelCase__ ) -> int: a : Optional[Any] = self.get_input_output_texts(A__ ) a : int = tokenizer.encode(A__ , add_special_tokens=A__ ) a : List[Any] = tokenizer.decode(A__ , clean_up_tokenization_spaces=A__ ) return text, ids def __a ( self ) -> int: pass # TODO add if relevant def __a ( self ) -> Any: pass # TODO add if relevant def __a ( self ) -> Union[str, Any]: pass # TODO add if relevant def __a ( self ) -> List[str]: a : Any = self.get_tokenizer() # Testing tokenization a : Optional[Any] = "こんにちは、世界。 こんばんは、㔺界。" a : Union[str, Any] = ["こん", "にちは", "、", "世界", "。", "<SP>", "こん", "ばんは", "、", "㔺界", "。"] a : Dict = tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens a : Optional[int] = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] a : Optional[int] = tokenizer.convert_tokens_to_ids(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens a : Optional[Any] = tokens + [tokenizer.unk_token] a : Tuple = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] a : List[str] = tokenizer.convert_tokens_to_ids(A__ ) self.assertListEqual(A__ , A__ ) def __a ( self ) -> Tuple: a : Any = self.get_tokenizer() # Testing tokenization a : Optional[int] = "こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。" a : Union[str, Any] = "こんにちは、、、、世界。こんばんは、、、、世界。" a : Any = tokenizer.encode(A__ ) a : Dict = tokenizer.decode(A__ ) self.assertEqual(A__ , A__ ) @slow def __a ( self ) -> List[Any]: a : Optional[int] = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) # Testing tokenization a : Dict = "こんにちは、世界。" a : List[Any] = "こんばんは、㔺界。😀" a : List[str] = "こんにちは、世界。こんばんは、世界。😀" a : Optional[Any] = tokenizer.encode(prefix_text + input_text ) a : Any = tokenizer.encode("" , prefix_text=prefix_text + input_text ) a : Union[str, Any] = tokenizer.encode(A__ , prefix_text=A__ ) a : str = tokenizer.decode(A__ ) a : Dict = tokenizer.decode(A__ ) a : Optional[Any] = tokenizer.decode(A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) self.assertEqual(A__ , A__ ) @slow def __a ( self ) -> str: a : Tuple = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) # Testing tokenization a : Optional[Any] = "こんにちは、世界。" a : int = "こんばんは、㔺界。😀" a : Tuple = len(tokenizer.encode(A__ ) ) - 2 a : Union[str, Any] = len(tokenizer.encode(A__ ) ) - 2 a : Any = [1] + [0] * (len_prefix + len_text + 1) a : Any = [1] * (len_prefix + len_text + 1) + [0] a : str = [1] + [1] * (len_prefix) + [0] * (len_text + 1) a : Any = tokenizer(prefix_text + input_text ).token_type_ids a : List[str] = tokenizer("" , prefix_text=prefix_text + input_text ).token_type_ids a : int = tokenizer(A__ , prefix_text=A__ ).token_type_ids self.assertListEqual(A__ , A__ ) self.assertListEqual(A__ , A__ ) self.assertListEqual(A__ , A__ ) @slow def __a ( self ) -> str: a : Tuple = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) a : int = tokenizer.encode("あンいワ" ) a : str = tokenizer.encode("" , prefix_text="あンいワ" ) a : Tuple = tokenizer.encode("いワ" , prefix_text="あン" ) self.assertEqual(tokenizer.decode(A__ ) , tokenizer.decode(A__ ) ) self.assertEqual(tokenizer.decode(A__ ) , tokenizer.decode(A__ ) ) self.assertNotEqual(A__ , A__ ) self.assertNotEqual(A__ , A__ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def __a ( self ) -> Tuple: a : str = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) a : Tuple = [["武田信玄", "は、"], ["織田信長", "の配下の、"]] a : List[str] = tokenizer(A__ , padding=A__ ) a : str = tokenizer.batch_encode_plus(A__ , padding=A__ ) # fmt: off a : Union[str, Any] = [[3_5993, 8640, 2_5948, 3_5998, 3_0647, 3_5675, 3_5999, 3_5999], [3_5993, 1_0382, 9868, 3_5998, 3_0646, 9459, 3_0646, 3_5675]] a : Union[str, Any] = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] a : Union[str, Any] = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , A__ ) self.assertListEqual(x_token.token_type_ids , A__ ) self.assertListEqual(x_token.attention_mask , A__ ) self.assertListEqual(x_token_a.input_ids , A__ ) self.assertListEqual(x_token_a.token_type_ids , A__ ) self.assertListEqual(x_token_a.attention_mask , A__ ) def __a ( self ) -> List[Any]: pass def __a ( self ) -> Optional[Any]: pass
633
from ...configuration_utils import PretrainedConfig class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = "bert-generation" def __init__( self : Optional[int] , A__ : List[Any]=5_03_58 , A__ : Any=10_24 , A__ : Any=24 , A__ : List[Any]=16 , A__ : List[Any]=40_96 , A__ : int="gelu" , A__ : List[str]=0.1 , A__ : List[str]=0.1 , A__ : str=5_12 , A__ : int=0.02 , A__ : Any=1E-12 , A__ : Optional[Any]=0 , A__ : List[str]=2 , A__ : Optional[int]=1 , A__ : str="absolute" , A__ : Any=True , **A__ : Optional[Any] , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , **A__ ) snake_case_ : str = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Optional[Any] = hidden_act snake_case_ : Tuple = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : Optional[Any] = initializer_range snake_case_ : Optional[int] = layer_norm_eps snake_case_ : str = position_embedding_type snake_case_ : Dict = use_cache
666
0
def a__ ( A__, A__ ): def get_matched_characters(A__, A__ ) -> str: SCREAMING_SNAKE_CASE_ : Tuple = [] SCREAMING_SNAKE_CASE_ : Tuple = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): SCREAMING_SNAKE_CASE_ : str = int(max(0, i - limit ) ) SCREAMING_SNAKE_CASE_ : Optional[int] = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE_ : List[Any] = F'''{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}''' return "".join(lowerCAmelCase_ ) # matching characters SCREAMING_SNAKE_CASE_ : List[Any] = get_matched_characters(lowerCAmelCase_, lowerCAmelCase_ ) SCREAMING_SNAKE_CASE_ : int = get_matched_characters(lowerCAmelCase_, lowerCAmelCase_ ) SCREAMING_SNAKE_CASE_ : Optional[int] = len(lowerCAmelCase_ ) # transposition SCREAMING_SNAKE_CASE_ : List[str] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_, lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: SCREAMING_SNAKE_CASE_ : str = 0.0 else: SCREAMING_SNAKE_CASE_ : Optional[Any] = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
101
import math def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [] snake_case_ : List[str] = 2 snake_case_ : Optional[int] = int(math.sqrt(lowerCAmelCase_ ) ) # Size of every segment snake_case_ : str = [True] * (end + 1) snake_case_ : Any = [] while start <= end: if temp[start] is True: in_prime.append(lowerCAmelCase_ ) for i in range(start * start , end + 1 , lowerCAmelCase_ ): snake_case_ : Union[str, Any] = False start += 1 prime += in_prime snake_case_ : Dict = end + 1 snake_case_ : Dict = min(2 * end , lowerCAmelCase_ ) while low <= n: snake_case_ : Any = [True] * (high - low + 1) for each in in_prime: snake_case_ : Optional[Any] = math.floor(low / each ) * each if t < low: t += each for j in range(lowerCAmelCase_ , high + 1 , lowerCAmelCase_ ): snake_case_ : List[Any] = False for j in range(len(lowerCAmelCase_ ) ): if temp[j] is True: prime.append(j + low ) snake_case_ : int = high + 1 snake_case_ : Union[str, Any] = min(high + end , lowerCAmelCase_ ) return prime print(sieve(1_0**6))
666
0
'''simple docstring''' def snake_case ( a_ : int ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : Union[str, Any] = int(lowerCAmelCase_ ) if decimal in (0, 1): # Exit cases for the recursion return str(lowerCAmelCase_ ) UpperCamelCase_ : Tuple = divmod(lowerCAmelCase_ , 2 ) return binary_recursive(lowerCAmelCase_ ) + str(lowerCAmelCase_ ) def snake_case ( a_ : str ) -> Tuple: """simple docstring""" UpperCamelCase_ : Dict = str(lowerCAmelCase_ ).strip() if not number: raise ValueError("""No input value was provided""" ) UpperCamelCase_ : Optional[int] = "-" if number.startswith("""-""" ) else "" UpperCamelCase_ : Any = number.lstrip("""-""" ) if not number.isnumeric(): raise ValueError("""Input value is not an integer""" ) return f"{negative}0b{binary_recursive(int(lowerCAmelCase_ ) )}" if __name__ == "__main__": from doctest import testmod testmod()
208
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self : List[str] , A__ : List[Any] , A__ : int=7 , A__ : Union[str, Any]=3 , A__ : List[str]=30 , A__ : Optional[int]=4_00 , A__ : Optional[Any]=True , A__ : Optional[int]=None , A__ : Optional[Any]=True , A__ : Any=[0.5, 0.5, 0.5] , A__ : int=[0.5, 0.5, 0.5] , A__ : Any=True , A__ : int=1 / 2_55 , A__ : List[str]=True , ) -> Dict: '''simple docstring''' snake_case_ : int = size if size is not None else {"shortest_edge": 18, "longest_edge": 13_33} snake_case_ : Any = parent snake_case_ : Optional[int] = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Union[str, Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Tuple = do_resize snake_case_ : Dict = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : List[Any] = image_std snake_case_ : Tuple = do_rescale snake_case_ : Any = rescale_factor snake_case_ : Optional[int] = do_pad def UpperCAmelCase__ ( self : int ) -> List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def UpperCAmelCase__ ( self : Optional[int] , A__ : Optional[int] , A__ : Any=False ) -> Optional[Any]: '''simple docstring''' if not batched: snake_case_ : Any = image_inputs[0] if isinstance(A__ , Image.Image ): snake_case_ ,snake_case_ : Dict = image.size else: snake_case_ ,snake_case_ : int = image.shape[1], image.shape[2] if w < h: snake_case_ : Dict = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[int] = self.size["shortest_edge"] elif w > h: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : str = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : Optional[int] = self.size["shortest_edge"] snake_case_ : List[Any] = self.size["shortest_edge"] else: snake_case_ : str = [] for image in image_inputs: snake_case_ ,snake_case_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : List[Any] = max(A__ , key=lambda A__ : item[0] )[0] snake_case_ : int = max(A__ , key=lambda A__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( _UpperCamelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[int] = ConditionalDetrImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : Tuple ) -> Dict: '''simple docstring''' snake_case_ : List[str] = ConditionalDetrImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Any ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A__ , "image_mean" ) ) self.assertTrue(hasattr(A__ , "image_std" ) ) self.assertTrue(hasattr(A__ , "do_normalize" ) ) self.assertTrue(hasattr(A__ , "do_resize" ) ) self.assertTrue(hasattr(A__ , "size" ) ) def UpperCAmelCase__ ( self : List[str] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 13_33} ) self.assertEqual(image_processor.do_pad , A__ ) snake_case_ : Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A__ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , A__ ) def UpperCAmelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' pass def UpperCAmelCase__ ( self : Dict ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ ) for image in image_inputs: self.assertIsInstance(A__ , Image.Image ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) snake_case_ : int = image_processing(A__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : int ) -> Any: '''simple docstring''' snake_case_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , numpify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , np.ndarray ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[str] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Optional[int] = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : Dict = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : Tuple ) -> str: '''simple docstring''' snake_case_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A__ , torchify=A__ ) for image in image_inputs: self.assertIsInstance(A__ , torch.Tensor ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(A__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Any = image_processing(A__ , return_tensors="pt" ).pixel_values snake_case_ ,snake_case_ : int = self.image_processor_tester.get_expected_values(A__ , batched=A__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def UpperCAmelCase__ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : Optional[Any] = json.loads(f.read() ) snake_case_ : int = {"image_id": 3_97_69, "annotations": target} # encode them snake_case_ : Optional[int] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) snake_case_ : Any = image_processing(images=A__ , annotations=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : List[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : List[str] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Tuple = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[Any] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Any = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify orig_size snake_case_ : Any = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : List[str] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) ) @slow def UpperCAmelCase__ ( self : int ) -> str: '''simple docstring''' snake_case_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : Any = json.loads(f.read() ) snake_case_ : Optional[Any] = {"file_name": "000000039769.png", "image_id": 3_97_69, "segments_info": target} snake_case_ : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor(format="coco_panoptic" ) snake_case_ : str = image_processing(images=A__ , annotations=A__ , masks_path=A__ , return_tensors="pt" ) # verify pixel values snake_case_ : int = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , A__ ) snake_case_ : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , A__ , atol=1E-4 ) ) # verify area snake_case_ : Optional[int] = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , A__ ) ) # verify boxes snake_case_ : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , A__ ) snake_case_ : str = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , A__ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , A__ ) ) # verify is_crowd snake_case_ : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , A__ ) ) # verify class_labels snake_case_ : Optional[int] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , A__ ) ) # verify masks snake_case_ : Union[str, Any] = 82_28_73 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , A__ ) # verify orig_size snake_case_ : Dict = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , A__ ) ) # verify size snake_case_ : str = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , A__ ) )
666
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowerCamelCase = { 'vocab_file': { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt' ), } } lowerCamelCase = { 'junnyu/roformer_chinese_small': 15_36, 'junnyu/roformer_chinese_base': 15_36, 'junnyu/roformer_chinese_char_small': 5_12, 'junnyu/roformer_chinese_char_base': 5_12, 'junnyu/roformer_small_discriminator': 1_28, 'junnyu/roformer_small_generator': 1_28, } lowerCamelCase = { 'junnyu/roformer_chinese_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_base': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_small': {'do_lower_case': True}, 'junnyu/roformer_chinese_char_base': {'do_lower_case': True}, 'junnyu/roformer_small_discriminator': {'do_lower_case': True}, 'junnyu/roformer_small_generator': {'do_lower_case': True}, } class A ( _UpperCamelCase ): UpperCamelCase__ : Dict =VOCAB_FILES_NAMES UpperCamelCase__ : int =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : str =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : Tuple =RoFormerTokenizer def __init__( self : Optional[int] , lowercase_ : List[Any]=None , lowercase_ : List[str]=None , lowercase_ : Optional[int]=True , lowercase_ : Any="[UNK]" , lowercase_ : int="[SEP]" , lowercase_ : Tuple="[PAD]" , lowercase_ : List[str]="[CLS]" , lowercase_ : Tuple="[MASK]" , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=None , **lowercase_ : Dict , ) -> Optional[int]: """simple docstring""" super().__init__( A__ , tokenizer_file=A__ , do_lower_case=A__ , unk_token=A__ , sep_token=A__ , pad_token=A__ , cls_token=A__ , mask_token=A__ , tokenize_chinese_chars=A__ , strip_accents=A__ , **A__ , ) _lowerCamelCase : List[str] =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get('lowercase' , A__ ) != do_lower_case or pre_tok_state.get('strip_accents' , A__ ) != strip_accents ): _lowerCamelCase : Dict =getattr(A__ , pre_tok_state.pop('type' ) ) _lowerCamelCase : str =do_lower_case _lowerCamelCase : List[Any] =strip_accents _lowerCamelCase : Dict =pre_tok_class(**A__ ) _lowerCamelCase : Optional[int] =do_lower_case def __getstate__( self : Tuple ) -> Tuple: """simple docstring""" _lowerCamelCase : int =self.__dict__.copy() _lowerCamelCase : Union[str, Any] =BertPreTokenizer() return state def __setstate__( self : Dict , lowercase_ : Tuple ) -> Any: """simple docstring""" _lowerCamelCase : List[str] =d _lowerCamelCase : List[str] =self.__dict__["_tokenizer"].get_vocab() _lowerCamelCase : Tuple =PreTokenizer.custom(JiebaPreTokenizer(A__ ) ) def lowerCamelCase ( self : List[Any] , lowercase_ : Tuple , lowercase_ : Any=None ) -> Optional[int]: """simple docstring""" _lowerCamelCase : Dict =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase ( self : Any , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ) -> List[int]: """simple docstring""" _lowerCamelCase : Any =[self.sep_token_id] _lowerCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self : Tuple , lowercase_ : str , lowercase_ : Optional[str] = None ) -> Tuple[str]: """simple docstring""" _lowerCamelCase : str =self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def lowerCamelCase ( self : Optional[Any] , lowercase_ : Any , lowercase_ : Dict=None , lowercase_ : Any=None , lowercase_ : List[Any]=False , **lowercase_ : Dict , ) -> str: """simple docstring""" _lowerCamelCase : Dict =BertPreTokenizer() return super().save_pretrained(A__ , A__ , A__ , A__ , **A__ )
464
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "Model type selected in the list: " + ", ".join(_UpperCamelCase )} ) _SCREAMING_SNAKE_CASE : str = field( default=_UpperCamelCase , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) _SCREAMING_SNAKE_CASE : int = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) _SCREAMING_SNAKE_CASE : int = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) _SCREAMING_SNAKE_CASE : bool = field( default=_UpperCamelCase , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) _SCREAMING_SNAKE_CASE : float = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) _SCREAMING_SNAKE_CASE : int = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) _SCREAMING_SNAKE_CASE : int = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Tuple = "train" _SCREAMING_SNAKE_CASE : Any = "dev" class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : SquadDataTrainingArguments _SCREAMING_SNAKE_CASE : List[SquadFeatures] _SCREAMING_SNAKE_CASE : Split _SCREAMING_SNAKE_CASE : bool def __init__( self : str , A__ : SquadDataTrainingArguments , A__ : PreTrainedTokenizer , A__ : Optional[int] = None , A__ : Union[str, Split] = Split.train , A__ : Optional[bool] = False , A__ : Optional[str] = None , A__ : Optional[str] = "pt" , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = args snake_case_ : int = is_language_sensitive snake_case_ : int = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(A__ , A__ ): try: snake_case_ : List[str] = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) snake_case_ : Tuple = mode # Load data features from cache or dataset file snake_case_ : Dict = "v2" if args.version_2_with_negative else "v1" snake_case_ : List[Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case_ : List[Any] = cached_features_file + ".lock" with FileLock(A__ ): if os.path.exists(A__ ) and not args.overwrite_cache: snake_case_ : int = time.time() snake_case_ : List[Any] = torch.load(A__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. snake_case_ : Tuple = self.old_features["features"] snake_case_ : List[str] = self.old_features.get("dataset" , A__ ) snake_case_ : Tuple = self.old_features.get("examples" , A__ ) logger.info( f"Loading features from cached file {cached_features_file} [took %.3f s]" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"Deleting cached file {cached_features_file} will allow dataset and examples to be cached in" " future run" ) else: if mode == Split.dev: snake_case_ : Tuple = self.processor.get_dev_examples(args.data_dir ) else: snake_case_ : Tuple = self.processor.get_train_examples(args.data_dir ) snake_case_ ,snake_case_ : Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=A__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=A__ , ) snake_case_ : Any = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , A__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]" ) def __len__( self : str ) -> Dict: '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[int] , A__ : Optional[int] ) -> Dict[str, torch.Tensor]: '''simple docstring''' snake_case_ : Any = self.features[i] snake_case_ : Optional[int] = torch.tensor(feature.input_ids , dtype=torch.long ) snake_case_ : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.token_type_ids , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.cls_index , dtype=torch.long ) snake_case_ : str = torch.tensor(feature.p_mask , dtype=torch.float ) snake_case_ : str = torch.tensor(feature.is_impossible , dtype=torch.float ) snake_case_ : Optional[int] = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: snake_case_ : Any = torch.tensor(feature.start_position , dtype=torch.long ) snake_case_ : List[Any] = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
666
0
"""simple docstring""" import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def __UpperCamelCase ( SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def __UpperCamelCase ( ) -> Dict: """simple docstring""" assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def __UpperCamelCase ( ) -> Optional[int]: """simple docstring""" __snake_case = "mock-s3-bucket" __snake_case = F'''s3://{mock_bucket}''' __snake_case = extract_path_from_uri(lowerCAmelCase_ ) assert dataset_path.startswith("s3://" ) is False __snake_case = "./local/path" __snake_case = extract_path_from_uri(lowerCAmelCase_ ) assert dataset_path == new_dataset_path def __UpperCamelCase ( SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" __snake_case = is_remote_filesystem(lowerCAmelCase_ ) assert is_remote is True __snake_case = fsspec.filesystem("file" ) __snake_case = is_remote_filesystem(lowerCAmelCase_ ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , lowerCAmelCase_ ) def __UpperCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" __snake_case = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __snake_case = input_paths[compression_fs_class.protocol] if input_path is None: __snake_case = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(lowerCAmelCase_ ) __snake_case = fsspec.filesystem(compression_fs_class.protocol , fo=lowerCAmelCase_ ) assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) __snake_case = os.path.basename(lowerCAmelCase_ ) __snake_case = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(lowerCAmelCase_ , "r" , encoding="utf-8" ) as f, open(lowerCAmelCase_ , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def __UpperCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" __snake_case = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __snake_case = compressed_file_paths[protocol] __snake_case = "dataset.jsonl" __snake_case = F'''{protocol}://{member_file_path}::{compressed_file_path}''' __snake_case = fsspec.get_fs_token_paths(lowerCAmelCase_ ) assert fs.isfile(lowerCAmelCase_ ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def __UpperCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" __snake_case = hf_api.dataset_info(lowerCAmelCase_ , token=lowerCAmelCase_ ) __snake_case = HfFileSystem(repo_info=lowerCAmelCase_ , token=lowerCAmelCase_ ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(lowerCAmelCase_ ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def __UpperCamelCase ( ) -> List[str]: """simple docstring""" __snake_case = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(lowerCAmelCase_ , lowerCAmelCase_ , clobber=lowerCAmelCase_ ) with pytest.warns(lowerCAmelCase_ ) as warning_info: importlib.reload(datasets.filesystems ) assert len(lowerCAmelCase_ ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
163
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { "microsoft/git-base": "https://huggingface.co/microsoft/git-base/resolve/main/config.json", } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "git_vision_model" def __init__( self : int , A__ : Union[str, Any]=7_68 , A__ : List[Any]=30_72 , A__ : Tuple=12 , A__ : Optional[Any]=12 , A__ : Optional[int]=3 , A__ : List[str]=2_24 , A__ : Dict=16 , A__ : int="quick_gelu" , A__ : Any=1E-5 , A__ : Tuple=0.0 , A__ : Optional[int]=0.02 , **A__ : List[str] , ) -> Optional[int]: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[Any] = hidden_size snake_case_ : str = intermediate_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = num_channels snake_case_ : Union[str, Any] = patch_size snake_case_ : List[str] = image_size snake_case_ : List[Any] = initializer_range snake_case_ : Any = attention_dropout snake_case_ : Any = layer_norm_eps snake_case_ : int = hidden_act @classmethod def UpperCAmelCase__ ( cls : List[Any] , A__ : Union[str, os.PathLike] , **A__ : Optional[int] ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(A__ ) snake_case_ ,snake_case_ : Tuple = cls.get_config_dict(A__ , **A__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": snake_case_ : Any = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(A__ , **A__ ) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Optional[Any] = "git" def __init__( self : Any , A__ : List[str]=None , A__ : List[str]=3_05_22 , A__ : Tuple=7_68 , A__ : Tuple=6 , A__ : str=12 , A__ : Any=30_72 , A__ : List[str]="gelu" , A__ : int=0.1 , A__ : Dict=0.1 , A__ : Any=10_24 , A__ : Optional[Any]=0.02 , A__ : Optional[Any]=1E-12 , A__ : Dict=0 , A__ : Any="absolute" , A__ : Tuple=True , A__ : Any=False , A__ : Tuple=1_01 , A__ : Tuple=1_02 , A__ : List[Any]=None , **A__ : List[str] , ) -> int: '''simple docstring''' super().__init__(bos_token_id=A__ , eos_token_id=A__ , pad_token_id=A__ , **A__ ) if vision_config is None: snake_case_ : int = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) snake_case_ : str = GitVisionConfig(**A__ ) snake_case_ : int = vocab_size snake_case_ : List[Any] = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : List[Any] = num_attention_heads snake_case_ : Any = hidden_act snake_case_ : Dict = intermediate_size snake_case_ : Any = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = initializer_range snake_case_ : List[str] = layer_norm_eps snake_case_ : Any = position_embedding_type snake_case_ : Union[str, Any] = use_cache snake_case_ : str = tie_word_embeddings snake_case_ : List[Any] = num_image_with_embedding snake_case_ : Dict = bos_token_id snake_case_ : int = eos_token_id def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' snake_case_ : Tuple = copy.deepcopy(self.__dict__ ) snake_case_ : Optional[int] = self.vision_config.to_dict() snake_case_ : Tuple = self.__class__.model_type return output
666
0
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=_UpperCamelCase ) class _a ( _UpperCamelCase ): a_ : str = field(default='image-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) a_ : ClassVar[Features] = Features({'image': Image()} ) a_ : ClassVar[Features] = Features({'labels': ClassLabel} ) a_ : str = "image" a_ : str = "labels" def _UpperCamelCase ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Dict ): if self.label_column not in features: raise ValueError(F'Column {self.label_column} is not present in features.' ) if not isinstance(features[self.label_column] , A__ ): raise ValueError(F'Column {self.label_column} is not a ClassLabel.' ) lowerCamelCase__ = copy.deepcopy(self ) lowerCamelCase__ = self.label_schema.copy() lowerCamelCase__ = features[self.label_column] lowerCamelCase__ = label_schema return task_template @property def _UpperCamelCase ( self : Dict ): return { self.image_column: "image", self.label_column: "labels", }
510
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: str ): def get_matched_characters(lowerCAmelCase_: str , lowerCAmelCase_: str ) -> str: snake_case_ : Tuple = [] snake_case_ : Tuple = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): snake_case_ : str = int(max(0 , i - limit ) ) snake_case_ : Optional[int] = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) snake_case_ : List[Any] = f"{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}" return "".join(lowerCAmelCase_ ) # matching characters snake_case_ : List[Any] = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : int = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = len(lowerCAmelCase_ ) # transposition snake_case_ : List[str] = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: snake_case_ : str = 0.0 else: snake_case_ : Optional[Any] = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters snake_case_ : Optional[Any] = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("hello", "world"))
666
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __magic_name__ = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ 'ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoCBertForCausalLM', 'RoCBertForMaskedLM', 'RoCBertForMultipleChoice', 'RoCBertForPreTraining', 'RoCBertForQuestionAnswering', 'RoCBertForSequenceClassification', 'RoCBertForTokenClassification', 'RoCBertLayer', 'RoCBertModel', 'RoCBertPreTrainedModel', 'load_tf_weights_in_roc_bert', ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
665
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) set_seed(7_7_0) UpperCAmelCase = { "c_attn": "att_proj", "c_proj": "out_proj", "c_fc": "in_proj", "transformer.": "", "h.": "layers.", "ln_1": "layernorm_1", "ln_2": "layernorm_2", "ln_f": "layernorm_final", "wpe": "position_embeds_layer", "wte": "input_embeds_layer", } UpperCAmelCase = { "text_small": { "repo_id": "suno/bark", "file_name": "text.pt", }, "coarse_small": { "repo_id": "suno/bark", "file_name": "coarse.pt", }, "fine_small": { "repo_id": "suno/bark", "file_name": "fine.pt", }, "text": { "repo_id": "suno/bark", "file_name": "text_2.pt", }, "coarse": { "repo_id": "suno/bark", "file_name": "coarse_2.pt", }, "fine": { "repo_id": "suno/bark", "file_name": "fine_2.pt", }, } UpperCAmelCase = os.path.dirname(os.path.abspath(__file__)) UpperCAmelCase = os.path.join(os.path.expanduser("~"), ".cache") UpperCAmelCase = os.path.join(os.getenv("XDG_CACHE_HOME", default_cache_dir), "suno", "bark_v0") def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int , lowerCAmelCase_: List[str]=False ): snake_case_ : Union[str, Any] = model_type if use_small: key += "_small" return os.path.join(lowerCAmelCase_ , REMOTE_MODEL_PATHS[key]["file_name"] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: str , lowerCAmelCase_: List[str] ): os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) hf_hub_download(repo_id=lowerCAmelCase_ , filename=lowerCAmelCase_ , local_dir=lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Any , lowerCAmelCase_: Dict , lowerCAmelCase_: List[str]=False , lowerCAmelCase_: Dict="text" ): if model_type == "text": snake_case_ : int = BarkSemanticModel snake_case_ : str = BarkSemanticConfig snake_case_ : Optional[Any] = BarkSemanticGenerationConfig elif model_type == "coarse": snake_case_ : str = BarkCoarseModel snake_case_ : Optional[int] = BarkCoarseConfig snake_case_ : Any = BarkCoarseGenerationConfig elif model_type == "fine": snake_case_ : Optional[int] = BarkFineModel snake_case_ : Tuple = BarkFineConfig snake_case_ : List[str] = BarkFineGenerationConfig else: raise NotImplementedError() snake_case_ : Optional[Any] = f"{model_type}_small" if use_small else model_type snake_case_ : Any = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(lowerCAmelCase_ ): logger.info(f"{model_type} model not found, downloading into `{CACHE_DIR}`." ) _download(model_info["repo_id"] , model_info["file_name"] ) snake_case_ : Any = torch.load(lowerCAmelCase_ , map_location=lowerCAmelCase_ ) # this is a hack snake_case_ : Union[str, Any] = checkpoint["model_args"] if "input_vocab_size" not in model_args: snake_case_ : str = model_args["vocab_size"] snake_case_ : Union[str, Any] = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments snake_case_ : Union[str, Any] = model_args.pop("n_head" ) snake_case_ : int = model_args.pop("n_embd" ) snake_case_ : Any = model_args.pop("n_layer" ) snake_case_ : List[str] = ConfigClass(**checkpoint["model_args"] ) snake_case_ : Optional[Any] = ModelClass(config=lowerCAmelCase_ ) snake_case_ : Tuple = GenerationConfigClass() snake_case_ : List[str] = model_generation_config snake_case_ : Optional[int] = checkpoint["model"] # fixup checkpoint snake_case_ : Optional[int] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(lowerCAmelCase_ ): # replace part of the key with corresponding layer name in HF implementation snake_case_ : Tuple = k[len(lowerCAmelCase_ ) :] for old_layer_name in new_layer_name_dict: snake_case_ : int = new_k.replace(lowerCAmelCase_ , new_layer_name_dict[old_layer_name] ) snake_case_ : int = state_dict.pop(lowerCAmelCase_ ) snake_case_ : Optional[int] = set(state_dict.keys() ) - set(model.state_dict().keys() ) snake_case_ : str = {k for k in extra_keys if not k.endswith(".attn.bias" )} snake_case_ : Any = set(model.state_dict().keys() ) - set(state_dict.keys() ) snake_case_ : List[Any] = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(lowerCAmelCase_ ) != 0: raise ValueError(f"extra keys found: {extra_keys}" ) if len(lowerCAmelCase_ ) != 0: raise ValueError(f"missing keys: {missing_keys}" ) model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) snake_case_ : str = model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) snake_case_ : Union[str, Any] = checkpoint["best_val_loss"].item() logger.info(f"model loaded: {round(n_params/1e6 , 1 )}M params, {round(lowerCAmelCase_ , 3 )} loss" ) model.eval() model.to(lowerCAmelCase_ ) del checkpoint, state_dict return model def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: List[Any] , lowerCAmelCase_: str=False , lowerCAmelCase_: int="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() snake_case_ : int = "cpu" # do conversion on cpu snake_case_ : Optional[Any] = _get_ckpt_path(lowerCAmelCase_ , use_small=lowerCAmelCase_ ) snake_case_ : Tuple = _load_model(lowerCAmelCase_ , lowerCAmelCase_ , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) # load bark initial model snake_case_ : int = _bark_load_model(lowerCAmelCase_ , "cpu" , model_type=lowerCAmelCase_ , use_small=lowerCAmelCase_ ) if model_type == "text": snake_case_ : Union[str, Any] = bark_model["model"] if model.num_parameters(exclude_embeddings=lowerCAmelCase_ ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model snake_case_ : Optional[Any] = 5 snake_case_ : Optional[int] = 1_0 if model_type in ["text", "coarse"]: snake_case_ : Optional[Any] = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) snake_case_ : str = bark_model(lowerCAmelCase_ )[0] snake_case_ : Tuple = model(lowerCAmelCase_ ) # take last logits snake_case_ : List[str] = output_new_model_total.logits[:, [-1], :] else: snake_case_ : Optional[int] = 3 snake_case_ : str = 8 snake_case_ : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) snake_case_ : Any = model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Union[str, Any] = bark_model(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[int] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: List[str] , lowerCAmelCase_: Any , lowerCAmelCase_: List[Any] , lowerCAmelCase_: int , lowerCAmelCase_: Optional[Any] , ): snake_case_ : Optional[Any] = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkSemanticConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = BarkCoarseConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[str] = BarkFineConfig.from_pretrained(os.path.join(lowerCAmelCase_ , "config.json" ) ) snake_case_ : List[Any] = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) snake_case_ : List[str] = BarkSemanticModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Optional[Any] = BarkCoarseModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Tuple = BarkFineModel.from_pretrained(lowerCAmelCase_ ) snake_case_ : Union[str, Any] = EncodecModel.from_pretrained("facebook/encodec_24khz" ) snake_case_ : Tuple = BarkConfig.from_sub_model_configs( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ : List[Any] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) snake_case_ : Optional[int] = BarkModel(lowerCAmelCase_ ) snake_case_ : int = semantic snake_case_ : List[str] = coarseAcoustic snake_case_ : str = fineAcoustic snake_case_ : Optional[Any] = codec snake_case_ : Any = bark_generation_config Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) bark.save_pretrained(lowerCAmelCase_ , repo_id=lowerCAmelCase_ , push_to_hub=lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("model_type", type=str, help="text, coarse or fine.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--is_small", action="store_true", help="convert the small version instead of the large.") UpperCAmelCase = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
666
0
'''simple docstring''' from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class _snake_case ( _UpperCamelCase ): def __init__( self , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = False , _lowerCamelCase = None , **_lowerCamelCase , ): UpperCAmelCase__ : Optional[int] = path_or_paths UpperCAmelCase__ : Any = split if split or isinstance(A__ , A__) else "train" UpperCAmelCase__ : List[str] = features UpperCAmelCase__ : Optional[int] = cache_dir UpperCAmelCase__ : Optional[int] = keep_in_memory UpperCAmelCase__ : Any = streaming UpperCAmelCase__ : Any = num_proc UpperCAmelCase__ : List[str] = kwargs @abstractmethod def snake_case__ ( self): pass class _snake_case ( _UpperCamelCase ): def __init__( self , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = False , _lowerCamelCase = False , _lowerCamelCase = None , **_lowerCamelCase , ): UpperCAmelCase__ : Any = features UpperCAmelCase__ : str = cache_dir UpperCAmelCase__ : Any = keep_in_memory UpperCAmelCase__ : int = streaming UpperCAmelCase__ : Union[str, Any] = num_proc UpperCAmelCase__ : Any = kwargs @abstractmethod def snake_case__ ( self): pass
407
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def __magic_name__ ( __lowerCAmelCase : List[Any] , __lowerCAmelCase : Dict=0.999 , __lowerCAmelCase : str="cosine" , ) -> Dict: if alpha_transform_type == "cosine": def alpha_bar_fn(__lowerCAmelCase : List[str] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__lowerCAmelCase : Optional[int] ): return math.exp(t * -12.0 ) else: raise ValueError(f'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) __lowerCamelCase = [] for i in range(lowerCAmelCase_ ): __lowerCamelCase = i / num_diffusion_timesteps __lowerCamelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(lowerCAmelCase_ ) / alpha_bar_fn(lowerCAmelCase_ ) , lowerCAmelCase_ ) ) return torch.tensor(lowerCAmelCase_ , dtype=torch.floataa ) class lowerCAmelCase__ ( _UpperCamelCase , _UpperCamelCase ): a__ : Union[str, Any] = [e.name for e in KarrasDiffusionSchedulers] a__ : Optional[Any] = 2 @register_to_config def __init__( self : List[Any] , SCREAMING_SNAKE_CASE__ : int = 10_00 , SCREAMING_SNAKE_CASE__ : float = 0.00085 , SCREAMING_SNAKE_CASE__ : float = 0.012 , SCREAMING_SNAKE_CASE__ : str = "linear" , SCREAMING_SNAKE_CASE__ : Optional[Union[np.ndarray, List[float]]] = None , SCREAMING_SNAKE_CASE__ : str = "epsilon" , SCREAMING_SNAKE_CASE__ : str = "linspace" , SCREAMING_SNAKE_CASE__ : int = 0 , ) -> Dict: if trained_betas is not None: __lowerCamelCase = torch.tensor(A__ , dtype=torch.floataa ) elif beta_schedule == "linear": __lowerCamelCase = torch.linspace(A__ , A__ , A__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __lowerCamelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , A__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __lowerCamelCase = betas_for_alpha_bar(A__ ) else: raise NotImplementedError(f'''{beta_schedule} does is not implemented for {self.__class__}''' ) __lowerCamelCase = 1.0 - self.betas __lowerCamelCase = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(A__ , A__ , A__ ) def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Tuple=None ) -> str: if schedule_timesteps is None: __lowerCamelCase = self.timesteps __lowerCamelCase = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: __lowerCamelCase = 1 if len(A__ ) > 1 else 0 else: __lowerCamelCase = timestep.cpu().item() if torch.is_tensor(A__ ) else timestep __lowerCamelCase = self._index_counter[timestep_int] return indices[pos].item() @property def __A ( self : Dict ) -> Optional[int]: if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def __A ( self : Dict , SCREAMING_SNAKE_CASE__ : torch.FloatTensor , SCREAMING_SNAKE_CASE__ : Union[float, torch.FloatTensor] , ) -> torch.FloatTensor: __lowerCamelCase = self.index_for_timestep(A__ ) if self.state_in_first_order: __lowerCamelCase = self.sigmas[step_index] else: __lowerCamelCase = self.sigmas_interpol[step_index] __lowerCamelCase = sample / ((sigma**2 + 1) ** 0.5) return sample def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, torch.device] = None , SCREAMING_SNAKE_CASE__ : Optional[int] = None , ) -> int: __lowerCamelCase = num_inference_steps __lowerCamelCase = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": __lowerCamelCase = np.linspace(0 , num_train_timesteps - 1 , A__ , dtype=A__ )[::-1].copy() elif self.config.timestep_spacing == "leading": __lowerCamelCase = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __lowerCamelCase = (np.arange(0 , A__ ) * step_ratio).round()[::-1].copy().astype(A__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": __lowerCamelCase = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __lowerCamelCase = (np.arange(A__ , 0 , -step_ratio )).round().copy().astype(A__ ) timesteps -= 1 else: raise ValueError( f'''{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.''' ) __lowerCamelCase = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) __lowerCamelCase = torch.from_numpy(np.log(A__ ) ).to(A__ ) __lowerCamelCase = np.interp(A__ , np.arange(0 , len(A__ ) ) , A__ ) __lowerCamelCase = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) __lowerCamelCase = torch.from_numpy(A__ ).to(device=A__ ) # interpolate sigmas __lowerCamelCase = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() __lowerCamelCase = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) __lowerCamelCase = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(A__ ).startswith('''mps''' ): # mps does not support float64 __lowerCamelCase = torch.from_numpy(A__ ).to(A__ , dtype=torch.floataa ) else: __lowerCamelCase = torch.from_numpy(A__ ).to(A__ ) # interpolate timesteps __lowerCamelCase = self.sigma_to_t(A__ ).to(A__ , dtype=timesteps.dtype ) __lowerCamelCase = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() __lowerCamelCase = torch.cat([timesteps[:1], interleaved_timesteps] ) __lowerCamelCase = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter __lowerCamelCase = defaultdict(A__ ) def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int ) -> str: __lowerCamelCase = sigma.log() # get distribution __lowerCamelCase = log_sigma - self.log_sigmas[:, None] # get sigmas range __lowerCamelCase = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) __lowerCamelCase = low_idx + 1 __lowerCamelCase = self.log_sigmas[low_idx] __lowerCamelCase = self.log_sigmas[high_idx] # interpolate sigmas __lowerCamelCase = (low - log_sigma) / (low - high) __lowerCamelCase = w.clamp(0 , 1 ) # transform interpolation to time range __lowerCamelCase = (1 - w) * low_idx + w * high_idx __lowerCamelCase = t.view(sigma.shape ) return t @property def __A ( self : int ) -> str: return self.sample is None def __A ( self : Tuple , SCREAMING_SNAKE_CASE__ : Union[torch.FloatTensor, np.ndarray] , SCREAMING_SNAKE_CASE__ : Union[float, torch.FloatTensor] , SCREAMING_SNAKE_CASE__ : Union[torch.FloatTensor, np.ndarray] , SCREAMING_SNAKE_CASE__ : bool = True , ) -> Union[SchedulerOutput, Tuple]: __lowerCamelCase = self.index_for_timestep(A__ ) # advance index counter by 1 __lowerCamelCase = timestep.cpu().item() if torch.is_tensor(A__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: __lowerCamelCase = self.sigmas[step_index] __lowerCamelCase = self.sigmas_interpol[step_index + 1] __lowerCamelCase = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method __lowerCamelCase = self.sigmas[step_index - 1] __lowerCamelCase = self.sigmas_interpol[step_index] __lowerCamelCase = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API __lowerCamelCase = 0 __lowerCamelCase = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": __lowerCamelCase = sigma_hat if self.state_in_first_order else sigma_interpol __lowerCamelCase = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": __lowerCamelCase = sigma_hat if self.state_in_first_order else sigma_interpol __lowerCamelCase = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError('''prediction_type not implemented yet: sample''' ) else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`''' ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order __lowerCamelCase = (sample - pred_original_sample) / sigma_hat # 3. delta timestep __lowerCamelCase = sigma_interpol - sigma_hat # store for 2nd order step __lowerCamelCase = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order __lowerCamelCase = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep __lowerCamelCase = sigma_next - sigma_hat __lowerCamelCase = self.sample __lowerCamelCase = None __lowerCamelCase = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=A__ ) def __A ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : torch.FloatTensor , SCREAMING_SNAKE_CASE__ : torch.FloatTensor , SCREAMING_SNAKE_CASE__ : torch.FloatTensor , ) -> torch.FloatTensor: __lowerCamelCase = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(A__ ): # mps does not support float64 __lowerCamelCase = self.timesteps.to(original_samples.device , dtype=torch.floataa ) __lowerCamelCase = timesteps.to(original_samples.device , dtype=torch.floataa ) else: __lowerCamelCase = self.timesteps.to(original_samples.device ) __lowerCamelCase = timesteps.to(original_samples.device ) __lowerCamelCase = [self.index_for_timestep(A__ , A__ ) for t in timesteps] __lowerCamelCase = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): __lowerCamelCase = sigma.unsqueeze(-1 ) __lowerCamelCase = original_samples + noise * sigma return noisy_samples def __len__( self : str ) -> List[Any]: return self.config.num_train_timesteps
298
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch UpperCAmelCase = logging.get_logger(__name__) class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : str = ["pixel_values"] def __init__( self : List[Any] , A__ : bool = True , A__ : Optional[Dict[str, int]] = None , A__ : PILImageResampling = PILImageResampling.BILINEAR , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , **A__ : int , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : Optional[int] = size if size is not None else {"shortest_edge": 2_56} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[str] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Any = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : int = do_resize snake_case_ : Optional[Any] = size snake_case_ : Optional[Any] = resample snake_case_ : Optional[int] = do_center_crop snake_case_ : List[Any] = crop_size snake_case_ : List[Any] = do_rescale snake_case_ : Optional[int] = rescale_factor snake_case_ : Optional[Any] = do_normalize snake_case_ : List[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : str , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : Any = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : int , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Union[str, Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Tuple = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : np.ndarray , A__ : float , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Tuple ) -> np.ndarray: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Dict , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , A__ : ImageInput , A__ : Optional[bool] = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : Dict[str, int] = None , A__ : Optional[bool] = None , A__ : Optional[float] = None , A__ : Optional[bool] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **A__ : Union[str, Any] , ) -> Optional[int]: '''simple docstring''' snake_case_ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Dict = size if size is not None else self.size snake_case_ : Optional[Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : str = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" ) snake_case_ : Dict = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Any = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : Dict = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Tuple = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Any = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : List[str] = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : Any = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : Union[str, Any] = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : Optional[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Any = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ ) def UpperCAmelCase__ ( self : List[str] , A__ : Dict , A__ : List[Tuple] = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(A__ ) != len(A__ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(A__ ): snake_case_ : Dict = target_sizes.numpy() snake_case_ : int = [] for idx in range(len(A__ ) ): snake_case_ : List[str] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=A__ ) snake_case_ : int = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(A__ ) else: snake_case_ : List[Any] = logits.argmax(dim=1 ) snake_case_ : List[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
666
0
from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
149
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase = { "configuration_blenderbot": [ "BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BlenderbotConfig", "BlenderbotOnnxConfig", ], "tokenization_blenderbot": ["BlenderbotTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["BlenderbotTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST", "BlenderbotForCausalLM", "BlenderbotForConditionalGeneration", "BlenderbotModel", "BlenderbotPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "TFBlenderbotForConditionalGeneration", "TFBlenderbotModel", "TFBlenderbotPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ "FlaxBlenderbotForConditionalGeneration", "FlaxBlenderbotModel", "FlaxBlenderbotPreTrainedModel", ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
666
0
import os import re import shutil import sys import tempfile import unittest import black UpperCAmelCase_ : str = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. UpperCAmelCase_ : Optional[int] = ''' \"\"\"\n Output class for the scheduler\'s step function output.\n\n Args:\n prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the\n denoising loop.\n pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images):\n The predicted denoised sample (x_{0}) based on the model output from the current timestep.\n `pred_original_sample` can be used to preview progress or for guidance.\n \"\"\"\n\n prev_sample: torch.FloatTensor\n pred_original_sample: Optional[torch.FloatTensor] = None\n''' class lowerCamelCase_ ( unittest.TestCase ): def lowerCAmelCase_ ( self : Tuple ): __A : Tuple = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir , """schedulers/""" ) ) __A : Optional[Any] = self.diffusers_dir shutil.copy( os.path.join(A__ , """src/diffusers/schedulers/scheduling_ddpm.py""" ) , os.path.join(self.diffusers_dir , """schedulers/scheduling_ddpm.py""" ) , ) def lowerCAmelCase_ ( self : Tuple ): __A : Optional[Any] = "src/diffusers" shutil.rmtree(self.diffusers_dir ) def lowerCAmelCase_ ( self : Optional[Any] , __A : Tuple , __A : Optional[Any] , __A : Optional[int] , __A : int=None ): __A : Any = comment + F"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: __A : List[str] = comment + F"""\nclass {class_name}(nn.Module):\n""" + overwrite_result __A : Any = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) __A : int = black.format_str(A__ , mode=A__ ) __A : Dict = os.path.join(self.diffusers_dir , """new_code.py""" ) with open(A__ , """w""" , newline="""\n""" ) as f: f.write(A__ ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(A__ ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=A__ ) with open(A__ , """r""" ) as f: self.assertTrue(f.read() , A__ ) def lowerCAmelCase_ ( self : Dict ): __A : Optional[int] = check_copies.find_code_in_diffusers("""schedulers.scheduling_ddpm.DDPMSchedulerOutput""" ) self.assertEqual(A__ , A__ ) def lowerCAmelCase_ ( self : Dict ): self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" , """DDPMSchedulerOutput""" , REFERENCE_CODE + """\n""" , ) # With no empty line at the end self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" , """DDPMSchedulerOutput""" , A__ , ) # Copy consistency with rename self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" , """TestSchedulerOutput""" , re.sub("""DDPM""" , """Test""" , A__ ) , ) # Copy consistency with a really long name __A : Any = "TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason" self.check_copy_consistency( F"""# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}""" , F"""{long_class_name}SchedulerOutput""" , re.sub("""Bert""" , A__ , A__ ) , ) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" , """TestSchedulerOutput""" , A__ , overwrite_result=re.sub("""DDPM""" , """Test""" , A__ ) , )
17
from ...configuration_utils import PretrainedConfig UpperCAmelCase = { "google/tapas-base-finetuned-sqa": ( "https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json" ), "google/tapas-base-finetuned-wtq": ( "https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json" ), "google/tapas-base-finetuned-wikisql-supervised": ( "https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json" ), "google/tapas-base-finetuned-tabfact": ( "https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json" ), } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = "tapas" def __init__( self : List[Any] , A__ : str=3_05_22 , A__ : Tuple=7_68 , A__ : List[Any]=12 , A__ : Optional[Any]=12 , A__ : Union[str, Any]=30_72 , A__ : Dict="gelu" , A__ : List[Any]=0.1 , A__ : str=0.1 , A__ : List[Any]=10_24 , A__ : Optional[int]=[3, 2_56, 2_56, 2, 2_56, 2_56, 10] , A__ : Union[str, Any]=0.02 , A__ : Tuple=1E-12 , A__ : Tuple=0 , A__ : Any=10.0 , A__ : List[str]=0 , A__ : List[str]=1.0 , A__ : Optional[Any]=None , A__ : Tuple=1.0 , A__ : Union[str, Any]=False , A__ : Any=None , A__ : Union[str, Any]=1.0 , A__ : int=1.0 , A__ : str=False , A__ : int=False , A__ : Optional[Any]="ratio" , A__ : str=None , A__ : int=None , A__ : Dict=64 , A__ : int=32 , A__ : Optional[Any]=False , A__ : List[str]=True , A__ : List[Any]=False , A__ : str=False , A__ : Any=True , A__ : Tuple=False , A__ : str=None , A__ : str=None , **A__ : List[str] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=A__ , **A__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) snake_case_ : int = vocab_size snake_case_ : int = hidden_size snake_case_ : Optional[Any] = num_hidden_layers snake_case_ : int = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[int] = intermediate_size snake_case_ : str = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Any = max_position_embeddings snake_case_ : List[Any] = type_vocab_sizes snake_case_ : str = initializer_range snake_case_ : Optional[Any] = layer_norm_eps # Fine-tuning task hyperparameters snake_case_ : Optional[int] = positive_label_weight snake_case_ : Dict = num_aggregation_labels snake_case_ : List[str] = aggregation_loss_weight snake_case_ : str = use_answer_as_supervision snake_case_ : int = answer_loss_importance snake_case_ : Any = use_normalized_answer_loss snake_case_ : int = huber_loss_delta snake_case_ : List[Any] = temperature snake_case_ : str = aggregation_temperature snake_case_ : List[str] = use_gumbel_for_cells snake_case_ : List[str] = use_gumbel_for_aggregation snake_case_ : Dict = average_approximation_function snake_case_ : List[str] = cell_selection_preference snake_case_ : Dict = answer_loss_cutoff snake_case_ : List[str] = max_num_rows snake_case_ : Union[str, Any] = max_num_columns snake_case_ : str = average_logits_per_cell snake_case_ : Union[str, Any] = select_one_column snake_case_ : Dict = allow_empty_column_selection snake_case_ : List[Any] = init_cell_selection_weights_to_zero snake_case_ : str = reset_position_index_per_cell snake_case_ : List[Any] = disable_per_token_loss # Aggregation hyperparameters snake_case_ : List[str] = aggregation_labels snake_case_ : Union[str, Any] = no_aggregation_label_index if isinstance(self.aggregation_labels , A__ ): snake_case_ : Optional[int] = {int(A__ ): v for k, v in aggregation_labels.items()}
666
0
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCamelCase ( _UpperCamelCase , unittest.TestCase ): lowerCamelCase : str =LongformerTokenizer lowerCamelCase : Tuple =True lowerCamelCase : List[Any] =LongformerTokenizerFast lowerCamelCase : Optional[int] =True def __a ( self ) -> List[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt a : Dict = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] a : str = dict(zip(A__ , range(len(A__ ) ) ) ) a : Any = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] a : str = {"unk_token": "<unk>"} a : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) a : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(A__ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(A__ ) ) def __a ( self , **lowerCAmelCase__ ) -> Tuple: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **A__ ) def __a ( self , **lowerCAmelCase__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **A__ ) def __a ( self , lowerCAmelCase__ ) -> str: a : List[str] = "lower newer" a : Optional[Any] = "lower newer" return input_text, output_text def __a ( self ) -> List[Any]: a : List[Any] = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) a : Optional[Any] = "lower newer" a : Tuple = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] a : Tuple = tokenizer.tokenize(A__ ) # , add_prefix_space=True) self.assertListEqual(A__ , A__ ) a : int = tokens + [tokenizer.unk_token] a : Any = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def __a ( self ) -> Union[str, Any]: a : List[str] = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" , add_special_tokens=A__ ) , [0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" , add_special_tokens=A__ ) , [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] , ) @slow def __a ( self ) -> int: a : Union[str, Any] = self.tokenizer_class.from_pretrained("allenai/longformer-base-4096" ) a : Dict = tokenizer.encode("sequence builders" , add_special_tokens=A__ ) a : Any = tokenizer.encode("multi-sequence build" , add_special_tokens=A__ ) a : Tuple = tokenizer.encode( "sequence builders" , add_special_tokens=A__ , add_prefix_space=A__ ) a : Tuple = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=A__ , add_prefix_space=A__ ) a : List[str] = tokenizer.build_inputs_with_special_tokens(A__ ) a : str = tokenizer.build_inputs_with_special_tokens(A__ , A__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def __a ( self ) -> List[Any]: a : int = self.get_tokenizer() a : Any = "Encode this sequence." a : Union[str, Any] = tokenizer.byte_encoder[" ".encode("utf-8" )[0]] # Testing encoder arguments a : Optional[int] = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) a : Optional[int] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(A__ , A__ ) a : Any = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) a : List[str] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(A__ , A__ ) tokenizer.add_special_tokens({"bos_token": "<s>"} ) a : List[Any] = tokenizer.encode(A__ , add_special_tokens=A__ ) a : Optional[int] = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(A__ , A__ ) # Testing spaces after special tokens a : str = "<mask>" tokenizer.add_special_tokens( {"mask_token": AddedToken(A__ , lstrip=A__ , rstrip=A__ )} ) # mask token has a left space a : str = tokenizer.convert_tokens_to_ids(A__ ) a : List[Any] = "Encode <mask> sequence" a : Dict = "Encode <mask>sequence" a : Dict = tokenizer.encode(A__ ) a : Union[str, Any] = encoded.index(A__ ) a : Tuple = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(A__ , A__ ) a : List[str] = tokenizer.encode(A__ ) a : List[str] = encoded.index(A__ ) a : Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(A__ , A__ ) def __a ( self ) -> Union[str, Any]: pass def __a ( self ) -> Optional[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): a : Any = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) a : Dict = self.tokenizer_class.from_pretrained(A__ , **A__ ) a : str = "A, <mask> AllenNLP sentence." a : str = tokenizer_r.encode_plus(A__ , add_special_tokens=A__ , return_token_type_ids=A__ ) a : str = tokenizer_p.encode_plus(A__ , add_special_tokens=A__ , return_token_type_ids=A__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) a : Optional[int] = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) a : List[Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( A__ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( A__ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) def __a ( self ) -> Tuple: for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): a : Optional[int] = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Dict = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) a : List[Any] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["add_prefix_space"] , A__ ) self.assertEqual(post_processor_state["add_prefix_space"] , A__ ) self.assertEqual(post_processor_state["trim_offsets"] , A__ ) def __a ( self ) -> str: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): a : List[Any] = "hello" # `hello` is a token in the vocabulary of `pretrained_name` a : Optional[int] = f"""{text_of_1_token} {text_of_1_token}""" a : List[Any] = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Tuple = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(A__ ) + 1, len(A__ ) + 1 + len(A__ )) , ) a : Dict = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Union[str, Any] = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(A__ ) + 1, len(A__ ) + 1 + len(A__ )) , ) a : Any = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Union[str, Any] = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(A__ ), len(A__ ) + 1 + len(A__ )) , ) a : Union[str, Any] = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Optional[int] = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(A__ ), len(A__ ) + 1 + len(A__ )) , ) a : Tuple = f""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) a : Optional[int] = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Optional[int] = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(A__ ) + 1, 1 + len(A__ ) + 1 + len(A__ )) , ) a : Optional[Any] = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : Dict = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(A__ ), 1 + len(A__ ) + 1 + len(A__ )) , ) a : int = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , add_prefix_space=A__ , trim_offsets=A__ ) a : str = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(A__ ), 1 + len(A__ ) + 1 + len(A__ )) , )
633
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : str , A__ : str ) -> Optional[int]: '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def UpperCAmelCase__ ( self : int , A__ : Optional[int] , A__ : Dict ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) class snake_case__ ( datasets.BeamBasedBuilder ): def UpperCAmelCase__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=A__ , ) def UpperCAmelCase__ ( self : Any , A__ : List[str] , A__ : str ) -> Optional[int]: '''simple docstring''' return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def UpperCAmelCase__ ( self : List[Any] , A__ : List[str] , A__ : Optional[int] ) -> List[str]: '''simple docstring''' import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(A__ ) def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def SCREAMING_SNAKE_CASE_ ( ): return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class snake_case__ ( _UpperCamelCase ): @require_beam def UpperCAmelCase__ ( self : str ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Dict = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[int] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : str ) -> Optional[Any]: '''simple docstring''' import apache_beam as beam snake_case_ : Tuple = beam.io.parquetio.WriteToParquet snake_case_ : Union[str, Any] = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[Any] = DummyBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: snake_case_ : int = partial(A__ , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) snake_case_ : Optional[Any] = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : Tuple = DummyBeamDataset(cache_dir=A__ ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def UpperCAmelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: snake_case_ : List[str] = NestedBeamDataset(cache_dir=A__ , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(A__ , builder.name , "default" , "0.0.0" , f"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) snake_case_ : int = builder.as_dataset() self.assertEqual(dset["train"].num_rows , A__ ) self.assertEqual(dset["train"].info.splits["train"].num_examples , A__ ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(A__ , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
666
0
import math def a__ ( A__, A__ = 0, A__ = 0 ): SCREAMING_SNAKE_CASE_ : Any = end or len(lowerCAmelCase_ ) for i in range(lowerCAmelCase_, lowerCAmelCase_ ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = i SCREAMING_SNAKE_CASE_ : Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: SCREAMING_SNAKE_CASE_ : Union[str, Any] = array[temp_index - 1] temp_index -= 1 SCREAMING_SNAKE_CASE_ : int = temp_index_value return array def a__ ( A__, A__, A__ ): # Max Heap SCREAMING_SNAKE_CASE_ : Any = index SCREAMING_SNAKE_CASE_ : int = 2 * index + 1 # Left Node SCREAMING_SNAKE_CASE_ : Any = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: SCREAMING_SNAKE_CASE_ : List[Any] = left_index if right_index < heap_size and array[largest] < array[right_index]: SCREAMING_SNAKE_CASE_ : Any = right_index if largest != index: SCREAMING_SNAKE_CASE_ : List[Any] = array[largest], array[index] heapify(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : List[Any] = len(lowerCAmelCase_ ) for i in range(n // 2, -1, -1 ): heapify(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) for i in range(n - 1, 0, -1 ): SCREAMING_SNAKE_CASE_ : str = array[0], array[i] heapify(lowerCAmelCase_, 0, lowerCAmelCase_ ) return array def a__ ( A__, A__, A__, A__ ): if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def a__ ( A__, A__, A__, A__ ): SCREAMING_SNAKE_CASE_ : int = low SCREAMING_SNAKE_CASE_ : Union[str, Any] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i SCREAMING_SNAKE_CASE_ : Dict = array[j], array[i] i += 1 def a__ ( A__ ): if len(lowerCAmelCase_ ) == 0: return array SCREAMING_SNAKE_CASE_ : Any = 2 * math.ceil(math.loga(len(lowerCAmelCase_ ) ) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 1_6 return intro_sort(lowerCAmelCase_, 0, len(lowerCAmelCase_ ), lowerCAmelCase_, lowerCAmelCase_ ) def a__ ( A__, A__, A__, A__, A__ ): while end - start > size_threshold: if max_depth == 0: return heap_sort(lowerCAmelCase_ ) max_depth -= 1 SCREAMING_SNAKE_CASE_ : Union[str, Any] = median_of_a(lowerCAmelCase_, lowerCAmelCase_, start + ((end - start) // 2) + 1, end - 1 ) SCREAMING_SNAKE_CASE_ : Optional[int] = partition(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) intro_sort(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = p return insertion_sort(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ : Optional[int] =input('Enter numbers separated by a comma : ').strip() lowerCAmelCase__ : int =[float(item) for item in user_input.split(',')] print(sort(unsorted))
101
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple ): class snake_case__ : def __init__( self : Any , A__ : List[str] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = metric_id class snake_case__ : _SCREAMING_SNAKE_CASE : List[str] = [MetricMock(_UpperCamelCase ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: Tuple , lowerCAmelCase_: int , lowerCAmelCase_: List[Any] , lowerCAmelCase_: Any , lowerCAmelCase_: List[str] ): if "tmp_path" in args: snake_case_ : List[Any] = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowerCAmelCase_ , match="https://huggingface.co/docs/evaluate" ): func(*lowerCAmelCase_ )
666
0
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A ( _UpperCamelCase, _UpperCamelCase, unittest.TestCase ): """simple docstring""" __a : str = IFInpaintingSuperResolutionPipeline __a : Dict = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} __a : Tuple = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) __a : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def _UpperCAmelCase ( self ): return self._get_superresolution_dummy_components() def _UpperCAmelCase ( self , __lowerCAmelCase , __lowerCAmelCase=0 ): if str(A__ ).startswith("""mps""" ): UpperCamelCase_ : Union[str, Any] = torch.manual_seed(A__ ) else: UpperCamelCase_ : Union[str, Any] = torch.Generator(device=A__ ).manual_seed(A__ ) UpperCamelCase_ : Optional[Any] = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) UpperCamelCase_ : List[str] = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) UpperCamelCase_ : List[Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) UpperCamelCase_ : List[str] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def _UpperCAmelCase ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _UpperCAmelCase ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def _UpperCAmelCase ( self ): super().test_save_load_floataa(expected_max_diff=1E-1 ) def _UpperCAmelCase ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _UpperCAmelCase ( self ): self._test_save_load_local() def _UpperCAmelCase ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
208
from __future__ import annotations import bisect def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Any = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: snake_case_ : Tuple = mid + 1 else: snake_case_ : Dict = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): if hi < 0: snake_case_ : Optional[Any] = len(lowerCAmelCase_ ) while lo < hi: snake_case_ : Union[str, Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: snake_case_ : Optional[Any] = mid + 1 else: snake_case_ : Tuple = mid return lo def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_left(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int = 0 , lowerCAmelCase_: int = -1 ): sorted_collection.insert(bisect_right(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Dict = 0 snake_case_ : Tuple = len(lowerCAmelCase_ ) - 1 while left <= right: snake_case_ : int = left + (right - left) // 2 snake_case_ : Optional[Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: snake_case_ : Optional[Any] = midpoint - 1 else: snake_case_ : Optional[int] = midpoint + 1 return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int ): snake_case_ : Optional[int] = bisect.bisect_left(lowerCAmelCase_ , lowerCAmelCase_ ) if index != len(lowerCAmelCase_ ) and sorted_collection[index] == item: return index return None def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int] , lowerCAmelCase_: int , lowerCAmelCase_: int , lowerCAmelCase_: int ): if right < left: return None snake_case_ : List[Any] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , midpoint - 1 ) else: return binary_search_by_recursion(lowerCAmelCase_ , lowerCAmelCase_ , midpoint + 1 , lowerCAmelCase_ ) if __name__ == "__main__": UpperCAmelCase = input("Enter numbers separated by comma:\n").strip() UpperCAmelCase = sorted(int(item) for item in user_input.split(",")) UpperCAmelCase = int(input("Enter a single number to be found in the list:\n")) UpperCAmelCase = binary_search(collection, target) if result is None: print(F"{target} was not found in {collection}.") else: print(F"{target} was found at position {result} in {collection}.")
666
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase = { 'configuration_upernet': ['UperNetConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ 'UperNetForSemanticSegmentation', 'UperNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
464
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , A__ : VQModel , A__ : UNetaDModel , A__ : DDIMScheduler ) -> List[Any]: '''simple docstring''' super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , A__ : int = 1 , A__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A__ : float = 0.0 , A__ : int = 50 , A__ : Optional[str] = "pil" , A__ : bool = True , **A__ : Optional[Any] , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' snake_case_ : Optional[int] = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) snake_case_ : List[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler snake_case_ : Any = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature snake_case_ : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case_ : List[Any] = {} if accepts_eta: snake_case_ : int = eta for t in self.progress_bar(self.scheduler.timesteps ): snake_case_ : Union[str, Any] = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual snake_case_ : Dict = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case_ : Union[str, Any] = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE snake_case_ : int = self.vqvae.decode(A__ ).sample snake_case_ : Dict = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ : Optional[int] = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
666
0
"""simple docstring""" import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __magic_name__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , snake_case_ : VQModel , snake_case_ : UNetaDModel , snake_case_ : DDIMScheduler ): super().__init__() self.register_modules(vqvae=A__ , unet=A__ , scheduler=A__ ) @torch.no_grad() def __call__( self : str , snake_case_ : int = 1 , snake_case_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , snake_case_ : float = 0.0 , snake_case_ : int = 50 , snake_case_ : Optional[str] = "pil" , snake_case_ : bool = True , **snake_case_ : Optional[Any] , ): __snake_case = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=A__ , ) __snake_case = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __snake_case = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A__ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature __snake_case = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case = {} if accepts_eta: __snake_case = eta for t in self.progress_bar(self.scheduler.timesteps ): __snake_case = self.scheduler.scale_model_input(A__ , A__ ) # predict the noise residual __snake_case = self.unet(A__ , A__ ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case = self.scheduler.step(A__ , A__ , A__ , **A__ ).prev_sample # decode the image latents with the VAE __snake_case = self.vqvae.decode(A__ ).sample __snake_case = (image / 2 + 0.5).clamp(0 , 1 ) __snake_case = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case = self.numpy_to_pil(A__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A__ )
163
from decimal import Decimal, getcontext from math import ceil, factorial def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) snake_case_ : List[str] = precision snake_case_ : Union[str, Any] = ceil(precision / 1_4 ) snake_case_ : List[str] = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : str = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : str = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): snake_case_ : Tuple = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCAmelCase = 5_0 print(F"The first {n} digits of pi is: {pi(n)}")
666
0
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _snake_case = logging.get_logger(__name__) _snake_case = { "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/config.json", # See all BART models at https://huggingface.co/models?filter=bart } class _a ( _UpperCamelCase ): a_ : Optional[int] = "bart" a_ : Optional[Any] = ["past_key_values"] a_ : List[Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple=5_02_65 , SCREAMING_SNAKE_CASE__ : Dict=10_24 , SCREAMING_SNAKE_CASE__ : Optional[int]=12 , SCREAMING_SNAKE_CASE__ : Tuple=40_96 , SCREAMING_SNAKE_CASE__ : str=16 , SCREAMING_SNAKE_CASE__ : Any=12 , SCREAMING_SNAKE_CASE__ : List[str]=40_96 , SCREAMING_SNAKE_CASE__ : int=16 , SCREAMING_SNAKE_CASE__ : str=0.0 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.0 , SCREAMING_SNAKE_CASE__ : List[str]="gelu" , SCREAMING_SNAKE_CASE__ : Optional[Any]=10_24 , SCREAMING_SNAKE_CASE__ : Optional[int]=0.1 , SCREAMING_SNAKE_CASE__ : Optional[Any]=0.0 , SCREAMING_SNAKE_CASE__ : List[str]=0.0 , SCREAMING_SNAKE_CASE__ : List[Any]=0.02 , SCREAMING_SNAKE_CASE__ : Dict=0.0 , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Optional[int]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=3 , SCREAMING_SNAKE_CASE__ : List[Any]=1 , SCREAMING_SNAKE_CASE__ : str=0 , SCREAMING_SNAKE_CASE__ : List[Any]=2 , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=2 , SCREAMING_SNAKE_CASE__ : Dict=2 , **SCREAMING_SNAKE_CASE__ : Union[str, Any] , ): lowerCamelCase__ = vocab_size lowerCamelCase__ = max_position_embeddings lowerCamelCase__ = d_model lowerCamelCase__ = encoder_ffn_dim lowerCamelCase__ = encoder_layers lowerCamelCase__ = encoder_attention_heads lowerCamelCase__ = decoder_ffn_dim lowerCamelCase__ = decoder_layers lowerCamelCase__ = decoder_attention_heads lowerCamelCase__ = dropout lowerCamelCase__ = attention_dropout lowerCamelCase__ = activation_dropout lowerCamelCase__ = activation_function lowerCamelCase__ = init_std lowerCamelCase__ = encoder_layerdrop lowerCamelCase__ = decoder_layerdrop lowerCamelCase__ = classifier_dropout lowerCamelCase__ = use_cache lowerCamelCase__ = encoder_layers lowerCamelCase__ = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=A__ , pad_token_id=A__ , bos_token_id=A__ , eos_token_id=A__ , is_encoder_decoder=A__ , decoder_start_token_id=A__ , forced_eos_token_id=A__ , **A__ , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , A__ ): lowerCamelCase__ = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' ) class _a ( _UpperCamelCase ): @property def _UpperCamelCase ( self : Dict ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ = {0: "batch"} lowerCamelCase__ = {0: "batch", 1: "past_decoder_sequence + sequence"} else: lowerCamelCase__ = {0: "batch", 1: "decoder_sequence"} lowerCamelCase__ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(A__ , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCamelCase__ = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ = self.num_layers for i in range(A__ ): lowerCamelCase__ = {0: "batch", 2: "past_sequence + sequence"} lowerCamelCase__ = {0: "batch", 2: "past_sequence + sequence"} else: lowerCamelCase__ = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def _UpperCamelCase ( self : List[str] ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ = super().outputs else: lowerCamelCase__ = super(A__ , self ).outputs if self.use_past: lowerCamelCase__ = self.num_layers for i in range(A__ ): lowerCamelCase__ = {0: "batch", 2: "past_sequence + sequence"} lowerCamelCase__ = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def _UpperCamelCase ( self : Tuple , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[TensorType] = None , ): lowerCamelCase__ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A__ , A__ , A__ , A__ , A__ ) # Generate decoder inputs lowerCamelCase__ = seq_length if not self.use_past else 1 lowerCamelCase__ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A__ , A__ , A__ , A__ , A__ ) lowerCamelCase__ = {F'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowerCamelCase__ = dict(**A__ , **A__ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ = common_inputs["input_ids"].shape lowerCamelCase__ = common_inputs["decoder_input_ids"].shape[1] lowerCamelCase__ = self.num_attention_heads lowerCamelCase__ = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ = decoder_seq_length + 3 lowerCamelCase__ = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCamelCase__ = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(A__ , A__ )] , dim=1 ) lowerCamelCase__ = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCamelCase__ = self.num_layers lowerCamelCase__ = min(A__ , A__ ) lowerCamelCase__ = max(A__ , A__ ) - min_num_layers lowerCamelCase__ = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(A__ ): common_inputs["past_key_values"].append( ( torch.zeros(A__ ), torch.zeros(A__ ), torch.zeros(A__ ), torch.zeros(A__ ), ) ) # TODO: test this. lowerCamelCase__ = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(A__ , A__ ): common_inputs["past_key_values"].append((torch.zeros(A__ ), torch.zeros(A__ )) ) return common_inputs def _UpperCamelCase ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[TensorType] = None , ): lowerCamelCase__ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A__ , A__ , A__ , A__ , A__ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ = common_inputs["input_ids"].shape # Not using the same length for past_key_values lowerCamelCase__ = seqlen + 2 lowerCamelCase__ = self.num_layers lowerCamelCase__ = self.num_attention_heads lowerCamelCase__ = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ = common_inputs["attention_mask"].dtype lowerCamelCase__ = torch.cat( [common_inputs['attention_mask'], torch.ones(A__ , A__ , dtype=A__ )] , dim=1 ) lowerCamelCase__ = [ (torch.zeros(A__ ), torch.zeros(A__ )) for _ in range(A__ ) ] return common_inputs def _UpperCamelCase ( self : List[str] , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[TensorType] = None , ): lowerCamelCase__ = compute_effective_axis_dimension( A__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCamelCase__ = tokenizer.num_special_tokens_to_add(A__ ) lowerCamelCase__ = compute_effective_axis_dimension( A__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A__ ) # Generate dummy inputs according to compute batch and sequence lowerCamelCase__ = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCamelCase__ = dict(tokenizer(A__ , return_tensors=A__ ) ) return common_inputs def _UpperCamelCase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : PreTrainedTokenizer , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : int = -1 , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[TensorType] = None , ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A__ , batch_size=A__ , seq_length=A__ , is_pair=A__ , framework=A__ ) elif self.task == "causal-lm": lowerCamelCase__ = self._generate_dummy_inputs_for_causal_lm( A__ , batch_size=A__ , seq_length=A__ , is_pair=A__ , framework=A__ ) else: lowerCamelCase__ = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A__ , batch_size=A__ , seq_length=A__ , is_pair=A__ , framework=A__ ) return common_inputs def _UpperCamelCase ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str] ): if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ = super()._flatten_past_key_values_(A__ , A__ , A__ , A__ ) else: lowerCamelCase__ = super(A__ , self )._flatten_past_key_values_( A__ , A__ , A__ , A__ )
510
def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int = 1_0_0_0 ): snake_case_ ,snake_case_ : List[str] = 1, 1 snake_case_ : List[str] = 2 while True: snake_case_ : Tuple = 0 snake_case_ : Union[str, Any] = fa + fa snake_case_ ,snake_case_ : str = fa, f index += 1 for _ in str(lowerCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
666
0
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def lowerCamelCase ( lowerCamelCase : dict , lowerCamelCase : str , lowerCamelCase : set , lowerCamelCase : set , lowerCamelCase : dict , lowerCamelCase : dict , lowerCamelCase : PriorityQueue , lowerCamelCase : dict , lowerCamelCase : float | int , ): for nxt, d in graph[v]: if nxt in visited_forward: continue A_ : str = cst_fwd.get(lowerCAmelCase_ , np.inf) A_ : Dict = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt)) A_ : int = new_cost_f A_ : Optional[int] = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: A_ : Optional[Any] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def lowerCamelCase ( lowerCamelCase : str , lowerCamelCase : str , lowerCamelCase : dict , lowerCamelCase : dict): A_ : Dict = -1 A_ : str = set() A_ : List[Any] = set() A_ : Dict = {source: 0} A_ : Optional[int] = {destination: 0} A_ : str = {source: None} A_ : Tuple = {destination: None} A_ : PriorityQueue[Any] = PriorityQueue() A_ : PriorityQueue[Any] = PriorityQueue() A_ : Union[str, Any] = np.inf queue_forward.put((0, source)) queue_backward.put((0, destination)) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): A_ : str = queue_forward.get() visited_forward.add(lowerCAmelCase_) A_ : List[Any] = queue_backward.get() visited_backward.add(lowerCAmelCase_) A_ : Union[str, Any] = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) A_ : List[str] = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: A_ : Optional[Any] = shortest_distance return shortest_path_distance __magic_name__ = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } __magic_name__ = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
665
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[int | float] , lowerCAmelCase_: int , lowerCAmelCase_: int ): if len(lowerCAmelCase_ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(lowerCAmelCase_ ) or left < -len(lowerCAmelCase_ ) or right >= len(lowerCAmelCase_ ) or right < -len(lowerCAmelCase_ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case_ : List[Any] = (left + right) >> 1 # the middle snake_case_ : Dict = find_max(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # find max in range[left, mid] snake_case_ : int = find_max(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
666
0
'''simple docstring''' import math class _snake_case : def __init__( self , _lowerCamelCase=0): # a graph with Node 0,1,...,N-1 UpperCAmelCase__ : List[str] = n UpperCAmelCase__ : Dict = [ [math.inf for j in range(0 , A__)] for i in range(0 , A__) ] # adjacency matrix for weight UpperCAmelCase__ : int = [ [math.inf for j in range(0 , A__)] for i in range(0 , A__) ] # dp[i][j] stores minimum distance from i to j def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : Optional[Any] = w def snake_case__ ( self): for k in range(0 , self.n): for i in range(0 , self.n): for j in range(0 , self.n): UpperCAmelCase__ : int = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j]) def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): return self.dp[u][v] if __name__ == "__main__": __A =Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
407
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } UpperCAmelCase = { "roberta-base": 5_1_2, "roberta-large": 5_1_2, "roberta-large-mnli": 5_1_2, "distilroberta-base": 5_1_2, "roberta-base-openai-detector": 5_1_2, "roberta-large-openai-detector": 5_1_2, } class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = ["input_ids", "attention_mask"] _SCREAMING_SNAKE_CASE : List[str] = RobertaTokenizer def __init__( self : Optional[int] , A__ : List[Any]=None , A__ : Optional[int]=None , A__ : List[str]=None , A__ : Dict="replace" , A__ : List[str]="<s>" , A__ : Optional[Any]="</s>" , A__ : List[str]="</s>" , A__ : List[Any]="<s>" , A__ : int="<unk>" , A__ : int="<pad>" , A__ : List[Any]="<mask>" , A__ : Any=False , A__ : Optional[int]=True , **A__ : Union[str, Any] , ) -> int: '''simple docstring''' super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : List[Any] = getattr(A__ , pre_tok_state.pop("type" ) ) snake_case_ : Any = add_prefix_space snake_case_ : List[Any] = pre_tok_class(**A__ ) snake_case_ : Optional[int] = add_prefix_space snake_case_ : List[str] = "post_processor" snake_case_ : Tuple = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case_ : str = tuple(state["sep"] ) if "cls" in state: snake_case_ : Tuple = tuple(state["cls"] ) snake_case_ : Tuple = False if state.get("add_prefix_space" , A__ ) != add_prefix_space: snake_case_ : Optional[Any] = add_prefix_space snake_case_ : str = True if state.get("trim_offsets" , A__ ) != trim_offsets: snake_case_ : Optional[int] = trim_offsets snake_case_ : List[Any] = True if changes_to_apply: snake_case_ : int = getattr(A__ , state.pop("type" ) ) snake_case_ : List[Any] = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCAmelCase__ ( self : Optional[Any] ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def UpperCAmelCase__ ( self : Tuple , A__ : Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case_ : Any = value def UpperCAmelCase__ ( self : int , *A__ : Optional[Any] , **A__ : int ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[Any] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Union[str, Any] , *A__ : Any , **A__ : List[Any] ) -> BatchEncoding: '''simple docstring''' snake_case_ : Optional[int] = kwargs.get("is_split_into_words" , A__ ) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCAmelCase__ ( self : int , A__ : List[str] , A__ : Union[str, Any]=None ) -> Any: '''simple docstring''' snake_case_ : List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Dict , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : str = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
666
0
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") SCREAMING_SNAKE_CASE__ : Any = logging.getLogger(__name__) @dataclass class lowerCAmelCase__ : a__ : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) a__ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) a__ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) a__ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) a__ : bool = field( default=_UpperCamelCase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) a__ : str = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) a__ : bool = field( default=_UpperCamelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class lowerCAmelCase__ : a__ : Optional[str] = field(default=_UpperCamelCase , metadata={"""help""": """The input training data file (a text file)."""} ) a__ : Optional[str] = field( default=_UpperCamelCase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) a__ : bool = field( default=_UpperCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) a__ : Optional[int] = field( default=_UpperCamelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) a__ : Optional[int] = field( default=_UpperCamelCase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) a__ : bool = field( default=_UpperCamelCase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) a__ : Optional[int] = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) a__ : Optional[int] = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __A ( self : Tuple ) -> Any: if self.train_file is not None: __lowerCamelCase = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: __lowerCamelCase = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class lowerCAmelCase__ : a__ : PreTrainedTokenizerBase a__ : Union[bool, str, PaddingStrategy] = True a__ : Optional[int] = None a__ : Optional[int] = None def __call__( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any ) -> Optional[int]: __lowerCamelCase = "label" if "label" in features[0].keys() else "labels" __lowerCamelCase = [feature.pop(A__ ) for feature in features] __lowerCamelCase = len(A__ ) __lowerCamelCase = len(features[0]['''input_ids'''] ) __lowerCamelCase = [ [{k: v[i] for k, v in feature.items()} for i in range(A__ )] for feature in features ] __lowerCamelCase = list(chain(*A__ ) ) __lowerCamelCase = self.tokenizer.pad( A__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten __lowerCamelCase = {k: v.view(A__ , A__ , -1 ) for k, v in batch.items()} # Add back labels __lowerCamelCase = torch.tensor(A__ , dtype=torch.intaa ) return batch def __magic_name__ ( ) -> List[str]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __lowerCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __lowerCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __lowerCamelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , lowerCAmelCase_ , lowerCAmelCase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() __lowerCamelCase = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase_ ) datasets.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(f'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. __lowerCamelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __lowerCamelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: __lowerCamelCase = {} if data_args.train_file is not None: __lowerCamelCase = data_args.train_file if data_args.validation_file is not None: __lowerCamelCase = data_args.validation_file __lowerCamelCase = data_args.train_file.split('''.''' )[-1] __lowerCamelCase = load_dataset( lowerCAmelCase_ , data_files=lowerCAmelCase_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. __lowerCamelCase = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __lowerCamelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) __lowerCamelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) __lowerCamelCase = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. __lowerCamelCase = [f'''ending{i}''' for i in range(4 )] __lowerCamelCase = "sent1" __lowerCamelCase = "sent2" if data_args.max_seq_length is None: __lowerCamelCase = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) __lowerCamelCase = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' f'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) __lowerCamelCase = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(__lowerCAmelCase : Optional[int] ): __lowerCamelCase = [[context] * 4 for context in examples[context_name]] __lowerCamelCase = examples[question_header_name] __lowerCamelCase = [ [f'''{header} {examples[end][i]}''' for end in ending_names] for i, header in enumerate(lowerCAmelCase_ ) ] # Flatten out __lowerCamelCase = list(chain(*lowerCAmelCase_ ) ) __lowerCamelCase = list(chain(*lowerCAmelCase_ ) ) # Tokenize __lowerCamelCase = tokenizer( lowerCAmelCase_ , lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowerCAmelCase_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) __lowerCamelCase = raw_datasets["train"] if data_args.max_train_samples is not None: __lowerCamelCase = min(len(lowerCAmelCase_ ) , data_args.max_train_samples ) __lowerCamelCase = train_dataset.select(range(lowerCAmelCase_ ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): __lowerCamelCase = train_dataset.map( lowerCAmelCase_ , batched=lowerCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) __lowerCamelCase = raw_datasets["validation"] if data_args.max_eval_samples is not None: __lowerCamelCase = min(len(lowerCAmelCase_ ) , data_args.max_eval_samples ) __lowerCamelCase = eval_dataset.select(range(lowerCAmelCase_ ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): __lowerCamelCase = eval_dataset.map( lowerCAmelCase_ , batched=lowerCAmelCase_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator __lowerCamelCase = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowerCAmelCase_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(__lowerCAmelCase : Tuple ): __lowerCamelCase = eval_predictions __lowerCamelCase = np.argmax(lowerCAmelCase_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer __lowerCamelCase = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , compute_metrics=lowerCAmelCase_ , ) # Training if training_args.do_train: __lowerCamelCase = None if training_args.resume_from_checkpoint is not None: __lowerCamelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: __lowerCamelCase = last_checkpoint __lowerCamelCase = trainer.train(resume_from_checkpoint=lowerCAmelCase_ ) trainer.save_model() # Saves the tokenizer too for easy upload __lowerCamelCase = train_result.metrics __lowerCamelCase = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCAmelCase_ ) ) __lowerCamelCase = min(lowerCAmelCase_ , len(lowerCAmelCase_ ) ) trainer.log_metrics('''train''' , lowerCAmelCase_ ) trainer.save_metrics('''train''' , lowerCAmelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) __lowerCamelCase = trainer.evaluate() __lowerCamelCase = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCAmelCase_ ) __lowerCamelCase = min(lowerCAmelCase_ , len(lowerCAmelCase_ ) ) trainer.log_metrics('''eval''' , lowerCAmelCase_ ) trainer.save_metrics('''eval''' , lowerCAmelCase_ ) __lowerCamelCase = { "finetuned_from": model_args.model_name_or_path, "tasks": "multiple-choice", "dataset_tags": "swag", "dataset_args": "regular", "dataset": "SWAG", "language": "en", } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase_ ) else: trainer.create_model_card(**lowerCAmelCase_ ) def __magic_name__ ( __lowerCAmelCase : Optional[int] ) -> Tuple: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
298
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) snake_case_ : Any = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } snake_case_ : List[str] = model(A__ )["last_hidden_state"] snake_case_ : str = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , A__ ) # compare the actual values for a slice. snake_case_ : List[str] = tf.convert_to_tensor( [ [ [0.068_1762, 0.1089_4451, 0.0677_2504], [-0.0642_3668, 0.0236_6615, 0.0432_9344], [-0.0605_7295, 0.0997_4135, -0.0007_0584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
666
0
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def snake_case_ ( lowerCAmelCase_ : NDArray[floataa] , lowerCAmelCase_ : NDArray[floataa] , lowerCAmelCase_ : list[int] , lowerCAmelCase_ : int , ): __lowercase : Optional[int] = coefficient_matrix.shape __lowercase : str = constant_matrix.shape if rowsa != colsa: __lowercase : Dict = F"Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}" raise ValueError(lowerCAmelCase_ ) if colsa != 1: __lowercase : Union[str, Any] = F"Constant matrix must be nx1 but received {rowsa}x{colsa}" raise ValueError(lowerCAmelCase_ ) if rowsa != rowsa: __lowercase : str = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " F"received {rowsa}x{colsa} and {rowsa}x{colsa}" ) raise ValueError(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) != rowsa: __lowercase : List[str] = ( "Number of initial values must be equal to number of rows in coefficient " F"matrix but received {len(lowerCAmelCase_ )} and {rowsa}" ) raise ValueError(lowerCAmelCase_ ) if iterations <= 0: raise ValueError("""Iterations must be at least 1""" ) __lowercase : NDArray[floataa] = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) __lowercase : List[Any] = table.shape strictly_diagonally_dominant(lowerCAmelCase_ ) # Iterates the whole matrix for given number of times for _ in range(lowerCAmelCase_ ): __lowercase : str = [] for row in range(lowerCAmelCase_ ): __lowercase : Optional[Any] = 0 for col in range(lowerCAmelCase_ ): if col == row: __lowercase : str = table[row][col] elif col == cols - 1: __lowercase : List[str] = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] __lowercase : Any = (temp + val) / denom new_val.append(lowerCAmelCase_ ) __lowercase : int = new_val return [float(lowerCAmelCase_ ) for i in new_val] def snake_case_ ( lowerCAmelCase_ : NDArray[floataa] ): __lowercase : List[Any] = table.shape __lowercase : int = True for i in range(0 , lowerCAmelCase_ ): __lowercase : List[str] = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("""Coefficient matrix is not strictly diagonally dominant""" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
149
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( _UpperCamelCase ): _SCREAMING_SNAKE_CASE : Dict = ["pixel_values"] def __init__( self : Union[str, Any] , A__ : bool = True , A__ : Dict[str, int] = None , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : bool = True , A__ : Dict[str, int] = None , A__ : bool = True , A__ : Union[int, float] = 1 / 2_55 , A__ : bool = True , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = True , **A__ : Optional[int] , ) -> None: '''simple docstring''' super().__init__(**A__ ) snake_case_ : str = size if size is not None else {"shortest_edge": 2_24} snake_case_ : Union[str, Any] = get_size_dict(A__ , default_to_square=A__ ) snake_case_ : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} snake_case_ : Dict = get_size_dict(A__ , default_to_square=A__ , param_name="crop_size" ) snake_case_ : str = do_resize snake_case_ : str = size snake_case_ : Optional[Any] = resample snake_case_ : Any = do_center_crop snake_case_ : Any = crop_size snake_case_ : str = do_rescale snake_case_ : Optional[Any] = rescale_factor snake_case_ : int = do_normalize snake_case_ : Optional[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ : List[str] = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ : int = do_convert_rgb def UpperCAmelCase__ ( self : Optional[int] , A__ : np.ndarray , A__ : Dict[str, int] , A__ : PILImageResampling = PILImageResampling.BICUBIC , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(A__ , default_to_square=A__ ) if "shortest_edge" not in size: raise ValueError(f"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) snake_case_ : str = get_resize_output_image_size(A__ , size=size["shortest_edge"] , default_to_square=A__ ) return resize(A__ , size=A__ , resample=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Tuple , A__ : np.ndarray , A__ : Dict[str, int] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[Any] , ) -> np.ndarray: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(A__ ) if "height" not in size or "width" not in size: raise ValueError(f"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(A__ , size=(size["height"], size["width"]) , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Optional[Any] , A__ : np.ndarray , A__ : Union[int, float] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : List[str] , ) -> str: '''simple docstring''' return rescale(A__ , scale=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : Any , A__ : np.ndarray , A__ : Union[float, List[float]] , A__ : Union[float, List[float]] , A__ : Optional[Union[str, ChannelDimension]] = None , **A__ : Any , ) -> np.ndarray: '''simple docstring''' return normalize(A__ , mean=A__ , std=A__ , data_format=A__ , **A__ ) def UpperCAmelCase__ ( self : List[Any] , A__ : ImageInput , A__ : bool = None , A__ : Dict[str, int] = None , A__ : PILImageResampling = None , A__ : bool = None , A__ : int = None , A__ : bool = None , A__ : float = None , A__ : bool = None , A__ : Optional[Union[float, List[float]]] = None , A__ : Optional[Union[float, List[float]]] = None , A__ : bool = None , A__ : Optional[Union[str, TensorType]] = None , A__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **A__ : Optional[Any] , ) -> PIL.Image.Image: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : Union[str, Any] = size if size is not None else self.size snake_case_ : Any = get_size_dict(A__ , param_name="size" , default_to_square=A__ ) snake_case_ : Optional[int] = resample if resample is not None else self.resample snake_case_ : int = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Tuple = get_size_dict(A__ , param_name="crop_size" , default_to_square=A__ ) snake_case_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : str = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : List[Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : List[str] = image_std if image_std is not None else self.image_std snake_case_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ : List[Any] = make_list_of_images(A__ ) if not valid_images(A__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ : Dict = [convert_to_rgb(A__ ) for image in images] # All transformations expect numpy arrays. snake_case_ : Dict = [to_numpy_array(A__ ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=A__ , size=A__ , resample=A__ ) for image in images] if do_center_crop: snake_case_ : Tuple = [self.center_crop(image=A__ , size=A__ ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=A__ , scale=A__ ) for image in images] if do_normalize: snake_case_ : int = [self.normalize(image=A__ , mean=A__ , std=A__ ) for image in images] snake_case_ : List[Any] = [to_channel_dimension_format(A__ , A__ ) for image in images] snake_case_ : Tuple = {"pixel_values": images} return BatchFeature(data=A__ , tensor_type=A__ )
666
0
# flake8: noqa # Lint as: python3 UpperCAmelCase_ : Dict = [ '''VerificationMode''', '''Version''', '''disable_progress_bar''', '''enable_progress_bar''', '''is_progress_bar_enabled''', '''experimental''', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
17
from __future__ import annotations def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): snake_case_ ,snake_case_ : Dict = position snake_case_ : int = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] snake_case_ : Union[str, Any] = [] for position in positions: snake_case_ ,snake_case_ : Union[str, Any] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(lowerCAmelCase_ ) return permissible_positions def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] ): return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: list[list[int]] , lowerCAmelCase_: tuple[int, int] , lowerCAmelCase_: int ): if is_complete(lowerCAmelCase_ ): return True for position in get_valid_pos(lowerCAmelCase_ , len(lowerCAmelCase_ ) ): snake_case_ ,snake_case_ : Dict = position if board[y][x] == 0: snake_case_ : List[str] = curr + 1 if open_knight_tour_helper(lowerCAmelCase_ , lowerCAmelCase_ , curr + 1 ): return True snake_case_ : Dict = 0 return False def SCREAMING_SNAKE_CASE_ ( lowerCAmelCase_: int ): snake_case_ : Any = [[0 for i in range(lowerCAmelCase_ )] for j in range(lowerCAmelCase_ )] for i in range(lowerCAmelCase_ ): for j in range(lowerCAmelCase_ ): snake_case_ : Optional[Any] = 1 if open_knight_tour_helper(lowerCAmelCase_ , (i, j) , 1 ): return board snake_case_ : Dict = 0 snake_case_ : str = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
666
0