code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _snake_case ( _snake_case , _snake_case , unittest.TestCase): UpperCamelCase__ : Optional[Any] =IFImgaImgSuperResolutionPipeline UpperCamelCase__ : int =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} UpperCamelCase__ : List[str] =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""}) UpperCamelCase__ : str =PipelineTesterMixin.required_optional_params - {"""latents"""} def A__ ( self : Tuple ): return self._get_superresolution_dummy_components() def A__ ( self : Optional[int], __lowercase : Any, __lowercase : Union[str, Any]=0 ): if str(lowerCAmelCase__ ).startswith("mps" ): lowercase__ = torch.manual_seed(lowerCAmelCase__ ) else: lowercase__ = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowercase__ = floats_tensor((1, 3, 32, 32), rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowercase__ = floats_tensor((1, 3, 16, 16), rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowercase__ = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available(), reason="XFormers attention is only available with CUDA and `xformers` installed", ) def A__ ( self : Optional[Any] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def A__ ( self : str ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda", reason="float16 requires CUDA" ) def A__ ( self : str ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def A__ ( self : Any ): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def A__ ( self : Any ): self._test_save_load_local() def A__ ( self : Any ): self._test_inference_batch_single_identical( expected_max_diff=1e-2, )
701
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) lowercase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase__ = 1 if upper_limit > 0: lowercase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: lowercase_ = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F'The Catalan numbers from 0 through {N} are:') print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
37
0
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def A__ ( *__lowercase : Dict, **__lowercase : List[str] ): pass @is_pipeline_test @require_vision class _snake_case ( unittest.TestCase): @require_torch def A__ ( self : Dict ): lowercase__ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification", ) lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(UpperCAmelCase_, candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(UpperCAmelCase_ ), [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["A", "B", "C"], batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], ], ) @require_tf def A__ ( self : Union[str, Any] ): lowercase__ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification", framework="tf" ) lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(UpperCAmelCase_, candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["A", "B", "C"], batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], [ {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, {"score": 0.333, "label": ANY(UpperCAmelCase_ )}, ], ], ) @slow @require_torch def A__ ( self : Any ): lowercase__ = pipeline( task="zero-shot-image-classification", model="openai/clip-vit-base-patch32", ) # This is an image of 2 cats with remotes and no planes lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(UpperCAmelCase_, candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["cat", "plane", "remote"], batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5, ) @slow @require_tf def A__ ( self : int ): lowercase__ = pipeline( task="zero-shot-image-classification", model="openai/clip-vit-base-patch32", framework="tf" ) # This is an image of 2 cats with remotes and no planes lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(UpperCAmelCase_, candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["cat", "plane", "remote"], batch_size=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ ), [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5, )
702
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase_ = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin lowercase_ = get_tests_dir("""fixtures/test_sentencepiece_bpe_char.model""") @require_sentencepiece @require_tokenizers class _snake_case ( _SCREAMING_SNAKE_CASE , unittest.TestCase): UpperCamelCase__ =SpeechTaTokenizer UpperCamelCase__ =False UpperCamelCase__ =True def A__ ( self : Optional[Any] ): super().setUp() # We have a SentencePiece fixture for testing lowercase__ = SpeechTaTokenizer(A_ ) lowercase__ = AddedToken("<mask>", lstrip=A_, rstrip=A_ ) lowercase__ = mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self : Optional[Any], __lowercase : Optional[int] ): lowercase__ = "this is a test" lowercase__ = "this is a test" return input_text, output_text def A__ ( self : Any, __lowercase : List[str], __lowercase : Optional[Any]=False, __lowercase : Optional[Any]=20, __lowercase : Union[str, Any]=5 ): lowercase__ , lowercase__ = self.get_input_output_texts(A_ ) lowercase__ = tokenizer.encode(A_, add_special_tokens=A_ ) lowercase__ = tokenizer.decode(A_, clean_up_tokenization_spaces=A_ ) return text, ids def A__ ( self : int ): lowercase__ = "<pad>" lowercase__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ), A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ), A_ ) def A__ ( self : str ): lowercase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], "<s>" ) self.assertEqual(vocab_keys[1], "<pad>" ) self.assertEqual(vocab_keys[-4], "œ" ) self.assertEqual(vocab_keys[-2], "<mask>" ) self.assertEqual(vocab_keys[-1], "<ctc_blank>" ) self.assertEqual(len(A_ ), 81 ) def A__ ( self : Any ): self.assertEqual(self.get_tokenizer().vocab_size, 79 ) def A__ ( self : Optional[Any] ): lowercase__ = self.get_tokenizers(do_lower_case=A_ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): lowercase__ = tokenizer.vocab_size lowercase__ = len(A_ ) self.assertNotEqual(A_, 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) lowercase__ = ["aaaaa bbbbbb", "cccccccccdddddddd"] lowercase__ = tokenizer.add_tokens(A_ ) lowercase__ = tokenizer.vocab_size lowercase__ = len(A_ ) self.assertNotEqual(A_, 0 ) self.assertEqual(A_, A_ ) self.assertEqual(A_, len(A_ ) ) self.assertEqual(A_, all_size + len(A_ ) ) lowercase__ = tokenizer.encode("aaaaa bbbbbb low cccccccccdddddddd l", add_special_tokens=A_ ) self.assertGreaterEqual(len(A_ ), 4 ) self.assertGreater(tokens[0], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3], tokenizer.vocab_size - 1 ) lowercase__ = {"eos_token": ">>>>|||<||<<|<<", "pad_token": "<<<<<|||>|>>>>|>"} lowercase__ = tokenizer.add_special_tokens(A_ ) lowercase__ = tokenizer.vocab_size lowercase__ = len(A_ ) self.assertNotEqual(A_, 0 ) self.assertEqual(A_, A_ ) self.assertEqual(A_, len(A_ ) ) self.assertEqual(A_, all_size_a + len(A_ ) ) lowercase__ = tokenizer.encode( ">>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l", add_special_tokens=A_ ) self.assertGreaterEqual(len(A_ ), 6 ) self.assertGreater(tokens[0], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0], tokens[1] ) self.assertGreater(tokens[-3], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3], tokens[-4] ) self.assertEqual(tokens[0], tokenizer.eos_token_id ) self.assertEqual(tokens[-3], tokenizer.pad_token_id ) def A__ ( self : List[Any] ): pass def A__ ( self : Dict ): pass def A__ ( self : Dict ): lowercase__ = self.get_tokenizer() lowercase__ = tokenizer.tokenize("This is a test" ) # fmt: off self.assertListEqual(A_, [SPIECE_UNDERLINE, "T", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "a", SPIECE_UNDERLINE, "t", "e", "s", "t"] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ), [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6], ) lowercase__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( A_, [SPIECE_UNDERLINE, "I", SPIECE_UNDERLINE, "w", "a", "s", SPIECE_UNDERLINE, "b", "o", "r", "n", SPIECE_UNDERLINE, "i", "n", SPIECE_UNDERLINE, "92000", ",", SPIECE_UNDERLINE, "a", "n", "d", SPIECE_UNDERLINE, "t", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "f", "a", "l", "s", "é", "."] ) lowercase__ = tokenizer.convert_tokens_to_ids(A_ ) # fmt: off self.assertListEqual(A_, [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on lowercase__ = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_, [SPIECE_UNDERLINE, "I", SPIECE_UNDERLINE, "w", "a", "s", SPIECE_UNDERLINE, "b", "o", "r", "n", SPIECE_UNDERLINE, "i", "n", SPIECE_UNDERLINE, "<unk>", ",", SPIECE_UNDERLINE, "a", "n", "d", SPIECE_UNDERLINE, "t", "h", "i", "s", SPIECE_UNDERLINE, "i", "s", SPIECE_UNDERLINE, "f", "a", "l", "s", "é", "."] ) @slow def A__ ( self : Optional[int] ): lowercase__ = [ "Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides " "general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural " "Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained " "models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.", "BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly " "conditioning on both left and right context in all layers.", "The quick brown fox jumps over the lazy dog.", ] # fmt: off lowercase__ = { "input_ids": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], "attention_mask": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=A_, model_name="microsoft/speecht5_asr", revision="c5ef64c71905caeccde0e4462ef3f9077224c524", sequences=A_, )
703
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") lowercase__ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f'''bert/{name}''' def create_tf_var(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = session.run(SCREAMING_SNAKE_CASE_ ) print(f'''Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace("-" , "_" ) + ".ckpt" ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_=None ): lowercase__ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory in which to save tensorflow model" ) lowercase__ = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
37
0
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class _snake_case : def __init__( self : Any, __lowercase : List[Any], __lowercase : List[str]=99, __lowercase : Optional[int]=13, __lowercase : str=7, __lowercase : str=9, __lowercase : str=True, __lowercase : Optional[int]=True, __lowercase : str=False, __lowercase : Dict=32, __lowercase : Dict=5, __lowercase : int=4, __lowercase : Optional[int]=37, __lowercase : Tuple=8, __lowercase : int=0.1, __lowercase : List[str]=0.002, __lowercase : List[str]=1, __lowercase : Tuple=0, __lowercase : Optional[int]=0, __lowercase : Optional[int]=None, __lowercase : List[str]=None, ): lowercase__ = parent lowercase__ = batch_size lowercase__ = encoder_seq_length lowercase__ = decoder_seq_length # For common tests lowercase__ = self.decoder_seq_length lowercase__ = is_training lowercase__ = use_attention_mask lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = d_ff lowercase__ = relative_attention_num_buckets lowercase__ = dropout_rate lowercase__ = initializer_factor lowercase__ = eos_token_id lowercase__ = pad_token_id lowercase__ = decoder_start_token_id lowercase__ = None lowercase__ = decoder_layers def A__ ( self : Dict ): return TaConfig.from_pretrained("google/umt5-base" ) def A__ ( self : str, __lowercase : Dict, __lowercase : Any, __lowercase : List[str], __lowercase : Optional[int]=None, __lowercase : Union[str, Any]=None, __lowercase : Tuple=None, __lowercase : int=None, __lowercase : Optional[int]=None, ): if attention_mask is None: lowercase__ = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowercase__ = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowercase__ = torch.ones(config.num_hidden_layers, config.num_attention_heads, device=UpperCamelCase_ ) if decoder_head_mask is None: lowercase__ = torch.ones(config.num_decoder_layers, config.num_attention_heads, device=UpperCamelCase_ ) if cross_attn_head_mask is None: lowercase__ = torch.ones( config.num_decoder_layers, config.num_attention_heads, device=UpperCamelCase_ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A__ ( self : Union[str, Any] ): lowercase__ = ids_tensor([self.batch_size, self.encoder_seq_length], self.vocab_size ) lowercase__ = ids_tensor([self.batch_size, self.decoder_seq_length], self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowercase__ = input_ids.clamp(self.pad_token_id + 1 ) lowercase__ = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowercase__ = self.get_config() lowercase__ = config.num_attention_heads lowercase__ = self.prepare_inputs_dict(UpperCamelCase_, UpperCamelCase_, UpperCamelCase_ ) return config, input_dict def A__ ( self : List[str] ): lowercase__ = self.prepare_config_and_inputs() return config, inputs_dict def A__ ( self : List[str] ): return TaConfig( vocab_size=166, d_model=self.hidden_size, d_ff=self.d_ff, d_kv=self.hidden_size // self.num_attention_heads, num_layers=self.num_hidden_layers, num_decoder_layers=self.decoder_layers, num_heads=self.num_attention_heads, relative_attention_num_buckets=self.relative_attention_num_buckets, dropout_rate=self.dropout_rate, initializer_factor=self.initializer_factor, eos_token_id=self.eos_token_id, bos_token_id=self.pad_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.decoder_start_token_id, ) def A__ ( self : List[str] ): return TaConfig( vocab_size=self.vocab_size, d_model=self.hidden_size, d_ff=self.d_ff, d_kv=self.hidden_size // self.num_attention_heads, num_layers=self.num_hidden_layers, num_decoder_layers=self.decoder_layers, num_heads=self.num_attention_heads, relative_attention_num_buckets=self.relative_attention_num_buckets, dropout_rate=self.dropout_rate, initializer_factor=self.initializer_factor, eos_token_id=self.eos_token_id, bos_token_id=self.pad_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.decoder_start_token_id, ) def A__ ( self : Any, __lowercase : Optional[int], __lowercase : str, __lowercase : str, __lowercase : Optional[Any], __lowercase : Optional[Any], __lowercase : Tuple, ): lowercase__ = UMTaModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() lowercase__ = model( input_ids=UpperCamelCase_, decoder_input_ids=UpperCamelCase_, attention_mask=UpperCamelCase_, decoder_attention_mask=UpperCamelCase_, ) lowercase__ = model(input_ids=UpperCamelCase_, decoder_input_ids=UpperCamelCase_ ) lowercase__ = result.last_hidden_state lowercase__ = result.past_key_values lowercase__ = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size(), (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size(), (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCamelCase_ ), config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ), 4 ) def A__ ( self : str, __lowercase : Any, __lowercase : Any, __lowercase : List[str], __lowercase : List[str], __lowercase : str, __lowercase : Any, ): lowercase__ = UMTaModel(config=UpperCamelCase_ ).get_decoder().to(UpperCamelCase_ ).eval() # first forward pass lowercase__ = model(UpperCamelCase_, use_cache=UpperCamelCase_ ) lowercase__ = model(UpperCamelCase_ ) lowercase__ = model(UpperCamelCase_, use_cache=UpperCamelCase_ ) self.parent.assertTrue(len(UpperCamelCase_ ) == len(UpperCamelCase_ ) ) self.parent.assertTrue(len(UpperCamelCase_ ) == len(UpperCamelCase_ ) + 1 ) lowercase__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase__ = ids_tensor((self.batch_size, 1), config.vocab_size ) # append to next input_ids and lowercase__ = torch.cat([input_ids, next_tokens], dim=-1 ) lowercase__ = model(UpperCamelCase_ )['last_hidden_state'] lowercase__ = model(UpperCamelCase_, past_key_values=UpperCamelCase_ )['last_hidden_state'] # select random slice lowercase__ = ids_tensor((1,), output_from_past.shape[-1] ).item() lowercase__ = output_from_no_past[:, -1, random_slice_idx].detach() lowercase__ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCamelCase_, UpperCamelCase_, atol=1e-3 ) ) def A__ ( self : Union[str, Any], __lowercase : Any, __lowercase : Optional[int], ): lowercase__ = UMTaModel(config=UpperCamelCase_ ).to(UpperCamelCase_ ).half().eval() lowercase__ = model(**UpperCamelCase_ )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCamelCase_ ).any().item() ) @require_torch class _snake_case ( lowercase__ , lowercase__ , lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCamelCase__ : List[Any] =(UMTaForConditionalGeneration,) if is_torch_available() else () UpperCamelCase__ : Union[str, Any] =( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCamelCase__ : Optional[Any] =True UpperCamelCase__ : List[str] =False UpperCamelCase__ : Any =False UpperCamelCase__ : str =True UpperCamelCase__ : List[str] =True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCamelCase__ : List[str] =[0.8, 0.9] def A__ ( self : List[str] ): lowercase__ = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def A__ ( self : Union[str, Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs() lowercase__ = UMTaModel(config_and_inputs[0] ).to(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCamelCase_, (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]), F'''{tmpdirname}/t5_test.onnx''', export_params=UpperCamelCase_, opset_version=9, input_names=["input_ids", "decoder_input_ids"], ) @unittest.skipIf(torch_device == "cpu", "Cant do half precision" ) def A__ ( self : str ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCamelCase_ ) def A__ ( self : Tuple ): lowercase__ = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowercase__ = self.model_tester.prepare_config_and_inputs() lowercase__ = config_and_inputs[0] lowercase__ = UMTaForConditionalGeneration(UpperCamelCase_ ).eval() model.to(UpperCamelCase_ ) lowercase__ = { 'head_mask': torch.zeros(config.num_layers, config.num_heads, device=UpperCamelCase_ ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers, config.num_heads, device=UpperCamelCase_ ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers, config.num_heads, device=UpperCamelCase_ ), } for attn_name, (name, mask) in zip(UpperCamelCase_, head_masking.items() ): lowercase__ = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowercase__ = torch.ones( config.num_decoder_layers, config.num_heads, device=UpperCamelCase_ ) lowercase__ = model.generate( config_and_inputs[1]["input_ids"], num_beams=1, max_length=3, output_attentions=UpperCamelCase_, return_dict_in_generate=UpperCamelCase_, **UpperCamelCase_, ) # We check the state of decoder_attentions and cross_attentions just from the last step lowercase__ = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ), 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def A__ ( self : str ): pass @require_torch @require_sentencepiece @require_tokenizers class _snake_case ( unittest.TestCase): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def A__ ( self : List[str] ): lowercase__ = UMTaForConditionalGeneration.from_pretrained("google/umt5-small", return_dict=UpperCamelCase_ ).to(UpperCamelCase_ ) lowercase__ = AutoTokenizer.from_pretrained("google/umt5-small", use_fast=UpperCamelCase_, legacy=UpperCamelCase_ ) lowercase__ = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowercase__ = tokenizer(UpperCamelCase_, return_tensors="pt", padding=UpperCamelCase_ ).input_ids # fmt: off lowercase__ = torch.tensor( [ [ 3_8530, 21_0703, 25_6299, 1410, 25_6298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 2_5922, 25_6299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 1_9014, 1_0620, 758, 25_6299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 25_6299, 1_4869, 281, 301, 25_6298, 275, 11_9983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 25_6299, 1_4869, 281, 2234, 289, 2275, 333,6_1391, 289, 25_6298, 543, 25_6297, 16_8714, 329, 25_6296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCamelCase_, UpperCamelCase_ ) lowercase__ = model.generate(input_ids.to(UpperCamelCase_ ) ) lowercase__ = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowercase__ = tokenizer.batch_decode(UpperCamelCase_ ) self.assertEqual(UpperCamelCase_, UpperCamelCase_ )
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_timesformer""": ["""TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimesformerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimesformerModel""", """TimesformerForVideoClassification""", """TimesformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowercase_ = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP", "SpeechT5Config", "SpeechT5HifiGanConfig", ], "feature_extraction_speecht5": ["SpeechT5FeatureExtractor"], "processing_speecht5": ["SpeechT5Processor"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["SpeechT5Tokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST", "SpeechT5ForSpeechToText", "SpeechT5ForSpeechToSpeech", "SpeechT5ForTextToSpeech", "SpeechT5Model", "SpeechT5PreTrainedModel", "SpeechT5HifiGan", ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
705
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() lowercase_ = { """bart""": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), """bert""": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-base-cased-finetuned-mrpc""": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """dpr""": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), """gpt2""": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlnet""": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm""": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm-roberta""": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """transfo-xl""": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """openai-gpt""": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """roberta""": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """layoutlm""": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), """roberta-large-mnli""": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """camembert""": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """flaubert""": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert""": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert-base-distilled-squad""": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert""": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert-visual-feature-encoder""": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """ctrl""": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """albert""": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """t5""": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """electra""": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """wav2vec2""": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) lowercase__ = config_class.from_json_file(SCREAMING_SNAKE_CASE_ ) lowercase__ = True lowercase__ = True print(f'''Building TensorFlow model from configuration: {config}''' ) lowercase__ = model_class(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): lowercase__ = cached_file( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: lowercase__ = load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if compare_with_pt_model: lowercase__ = tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE_ ) # build the network lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" ) lowercase__ = pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE_ , config=SCREAMING_SNAKE_CASE_ , state_dict=SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): lowercase__ = pt_model(**pt_model.dummy_inputs ) lowercase__ = pto[0].numpy() lowercase__ = tfo[0].numpy() lowercase__ = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(SCREAMING_SNAKE_CASE_ , save_format="h5" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ): if args_model_type is None: lowercase__ = list(MODEL_CLASSES.keys() ) else: lowercase__ = [args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE_ , start=1 ): print("=" * 100 ) print(f''' Converting model type {j}/{len(SCREAMING_SNAKE_CASE_ )}: {model_type}''' ) print("=" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: lowercase__ = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: lowercase__ = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , start=1 ): print("-" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue lowercase__ = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE_ )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 100 ) if config_shortcut_name in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = config_shortcut_name if model_shortcut_name in aws_model_maps: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE_ ): lowercase__ = "converted_model" convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE_ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE_ , config_file=SCREAMING_SNAKE_CASE_ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE_ , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=SCREAMING_SNAKE_CASE_ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE_ ) os.remove(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_dump_path""", default=None, type=str, required=True, help="""Path to the output Tensorflow dump file.""" ) parser.add_argument( """--model_type""", default=None, type=str, help=( F'Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ' """convert all the models from AWS.""" ), ) parser.add_argument( """--pytorch_checkpoint_path""", default=None, type=str, help=( """Path to the PyTorch checkpoint path or shortcut name to download from AWS. """ """If not given, will download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--config_file""", default=None, type=str, help=( """The config json file corresponding to the pre-trained model. \n""" """This specifies the model architecture. If not given and """ """--pytorch_checkpoint_path is not given or is a shortcut name """ """use the configuration associated to the shortcut name on the AWS""" ), ) parser.add_argument( """--compare_with_pt_model""", action="""store_true""", help="""Compare Tensorflow and PyTorch model predictions.""" ) parser.add_argument( """--use_cached_models""", action="""store_true""", help="""Use cached models if possible instead of updating to latest checkpoint versions.""", ) parser.add_argument( """--remove_cached_files""", action="""store_true""", help="""Remove pytorch models after conversion (save memory when converting in batches).""", ) parser.add_argument("""--only_convert_finetuned_models""", action="""store_true""", help="""Only convert finetuned models.""") lowercase_ = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
37
0
from typing import Union import fire import torch from tqdm import tqdm def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = "cpu" , SCREAMING_SNAKE_CASE_ = None ): lowercase__ = torch.load(lowerCamelCase_ , map_location=lowerCamelCase_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(lowerCamelCase_ , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) lowercase__ = v.half() if save_path is None: # overwrite src_path lowercase__ = src_path torch.save(lowerCamelCase_ , lowerCamelCase_ ) if __name__ == "__main__": fire.Fire(convert)
706
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
0
from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING lowercase_ = logging.get_logger(__name__) @add_end_docstrings(UpperCAmelCase_) class _snake_case ( UpperCAmelCase_): def __init__( self : List[Any], *__lowercase : int, **__lowercase : List[Any] ): super().__init__(*_lowercase, **_lowercase ) requires_backends(self, "vision" ) self.check_model_type(_lowercase ) def __call__( self : List[str], __lowercase : Union[str, Any], **__lowercase : Union[str, Any] ): return super().__call__(_lowercase, **_lowercase ) def A__ ( self : List[str], **__lowercase : Any ): return {}, {}, {} def A__ ( self : Tuple, __lowercase : int ): lowercase__ = load_image(_lowercase ) lowercase__ = image.size lowercase__ = self.image_processor(images=_lowercase, return_tensors=self.framework ) return model_inputs def A__ ( self : Dict, __lowercase : int ): lowercase__ = self.model(**_lowercase ) return model_outputs def A__ ( self : Dict, __lowercase : Dict ): lowercase__ = model_outputs.predicted_depth lowercase__ = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ), size=self.image_size[::-1], mode="bicubic", align_corners=_lowercase ) lowercase__ = prediction.squeeze().cpu().numpy() lowercase__ = (output * 255 / np.max(_lowercase )).astype("uint8" ) lowercase__ = Image.fromarray(_lowercase ) lowercase__ = {} lowercase__ = predicted_depth lowercase__ = depth return output_dict
707
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : def __init__( self : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Union[str, Any], __lowercase : str, __lowercase : List[Any], __lowercase : List[str]=0.2, __lowercase : List[str]=0.2 ): lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = conva_get[:2] lowercase__ = conva_get[2] lowercase__ = size_pa lowercase__ = rate_w lowercase__ = rate_t lowercase__ = [ np.mat(-1 * np.random.rand(self.conva[0], self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self : Any, __lowercase : List[str] ): # save model dict with pickle lowercase__ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(__lowercase, "wb" ) as f: pickle.dump(__lowercase, __lowercase ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls : Dict, __lowercase : Union[str, Any] ): # read saved model with open(__lowercase, "rb" ) as f: lowercase__ = pickle.load(__lowercase ) # noqa: S301 lowercase__ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) lowercase__ = model_dic.get("size_pooling1" ) lowercase__ = model_dic.get("num_bp1" ) lowercase__ = model_dic.get("num_bp2" ) lowercase__ = model_dic.get("num_bp3" ) lowercase__ = model_dic.get("rate_weight" ) lowercase__ = model_dic.get("rate_thre" ) # create model instance lowercase__ = CNN(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) # modify model parameter lowercase__ = model_dic.get("w_conv1" ) lowercase__ = model_dic.get("wkj" ) lowercase__ = model_dic.get("vji" ) lowercase__ = model_dic.get("thre_conv1" ) lowercase__ = model_dic.get("thre_bp2" ) lowercase__ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self : str, __lowercase : List[Any] ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self : List[str], __lowercase : Optional[Any] ): return round(__lowercase, 3 ) def A__ ( self : Optional[Any], __lowercase : Dict, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : str ): # convolution process lowercase__ = convs[0] lowercase__ = convs[1] lowercase__ = np.shape(__lowercase )[0] # get the data slice of original image data, data_focus lowercase__ = [] for i_focus in range(0, size_data - size_conv + 1, __lowercase ): for j_focus in range(0, size_data - size_conv + 1, __lowercase ): lowercase__ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ = [] lowercase__ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__lowercase ): lowercase__ = [] for i_focus in range(len(__lowercase ) ): lowercase__ = ( np.sum(np.multiply(data_focus[i_focus], w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape( __lowercase, __lowercase ) data_featuremap.append(__lowercase ) # expanding the data slice to One dimenssion lowercase__ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__lowercase ) ) lowercase__ = np.asarray(__lowercase ) return focus_list, data_featuremap def A__ ( self : List[Any], __lowercase : Any, __lowercase : List[Any], __lowercase : Union[str, Any]="average_pool" ): # pooling process lowercase__ = len(featuremaps[0] ) lowercase__ = int(size_map / size_pooling ) lowercase__ = [] for i_map in range(len(__lowercase ) ): lowercase__ = featuremaps[i_map] lowercase__ = [] for i_focus in range(0, __lowercase, __lowercase ): for j_focus in range(0, __lowercase, __lowercase ): lowercase__ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape(__lowercase, __lowercase ) featuremap_pooled.append(__lowercase ) return featuremap_pooled def A__ ( self : str, __lowercase : Optional[Any] ): # expanding three dimension data to one dimension list lowercase__ = [] for i in range(len(__lowercase ) ): lowercase__ = np.shape(data[i] ) lowercase__ = data[i].reshape(1, shapes[0] * shapes[1] ) lowercase__ = data_listed.getA().tolist()[0] data_expanded.extend(__lowercase ) lowercase__ = np.asarray(__lowercase ) return data_expanded def A__ ( self : Optional[int], __lowercase : Optional[int] ): # expanding matrix to one dimension list lowercase__ = np.asarray(__lowercase ) lowercase__ = np.shape(__lowercase ) lowercase__ = data_mat.reshape(1, shapes[0] * shapes[1] ) return data_expanded def A__ ( self : str, __lowercase : Tuple, __lowercase : List[Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = [] lowercase__ = 0 for i_map in range(__lowercase ): lowercase__ = np.ones((size_map, size_map) ) for i in range(0, __lowercase, __lowercase ): for j in range(0, __lowercase, __lowercase ): lowercase__ = pd_pool[ i_pool ] lowercase__ = i_pool + 1 lowercase__ = np.multiply( __lowercase, np.multiply(out_map[i_map], (1 - out_map[i_map]) ) ) pd_all.append(__lowercase ) return pd_all def A__ ( self : Tuple, __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : List[str]=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(__lowercase )) ) print((" - - Shape: Teach_Data ", np.shape(__lowercase )) ) lowercase__ = 0 lowercase__ = [] lowercase__ = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase__ = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(__lowercase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ = np.asmatrix(datas_train[p] ) lowercase__ = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = np.shape(__lowercase ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = np.dot(__lowercase, self.vji.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = np.dot(__lowercase, self.wkj.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ = np.multiply( (data_teach - bp_outa), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.multiply( np.dot(__lowercase, self.wkj ), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.dot(__lowercase, self.vji ) lowercase__ = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ = pd_conva_pooled.T.getA().tolist() lowercase__ = self._calculate_gradient_from_pool( __lowercase, __lowercase, shape_featuremapa[0], shape_featuremapa[1], self.size_poolinga, ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ = self.rate_weight * np.dot(__lowercase, __lowercase ) lowercase__ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ = rp + 1 lowercase__ = error_count / patterns all_mse.append(__lowercase ) def draw_error(): lowercase__ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__lowercase, "+-" ) plt.plot(__lowercase, "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(__lowercase, alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self : List[str], __lowercase : Optional[int] ): # model predict lowercase__ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(__lowercase )) ) for p in range(len(__lowercase ) ): lowercase__ = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = bp_outa * self.vji.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = bp_outa * self.wkj.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ = [list(map(self.do_round, __lowercase ) ) for each in produce_out] return np.asarray(__lowercase ) def A__ ( self : int, __lowercase : Any ): # return the data of image after convoluting process so we can check it out lowercase__ = np.asmatrix(__lowercase ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
37
0
class _snake_case : def __init__( self : Dict, __lowercase : Any ): lowercase__ = val lowercase__ = None lowercase__ = None def A__ ( self : Any, __lowercase : str ): if self.val: if val < self.val: if self.left is None: lowercase__ = Node(__lowercase ) else: self.left.insert(__lowercase ) elif val > self.val: if self.right is None: lowercase__ = Node(__lowercase ) else: self.right.insert(__lowercase ) else: lowercase__ = val def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if root: inorder(root.left , __A ) res.append(root.val ) inorder(root.right , __A ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if len(__A ) == 0: return arr lowercase__ = Node(arr[0] ) for i in range(1 , len(__A ) ): root.insert(arr[i] ) # Traverse BST in order. lowercase__ = [] inorder(__A , __A ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
708
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=SCREAMING_SNAKE_CASE_ , num_labels=1000 , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): lowercase__ = get_config(SCREAMING_SNAKE_CASE_ ) # load original model from timm lowercase__ = create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=SCREAMING_SNAKE_CASE_ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) lowercase__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # verify logits with torch.no_grad(): lowercase__ = model(SCREAMING_SNAKE_CASE_ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(SCREAMING_SNAKE_CASE_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE_ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowercase_ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
37
0
from math import factorial lowercase_ = {str(digit): factorial(digit) for digit in range(10)} def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise TypeError("Parameter number must be int" ) if number < 0: raise ValueError("Parameter number must be greater than or equal to 0" ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(__lowerCAmelCase ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = 60 , SCREAMING_SNAKE_CASE_ = 100_0000 ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise TypeError("Parameters chain_length and number_limit must be int" ) if chain_length <= 0 or number_limit <= 0: raise ValueError( "Parameters chain_length and number_limit must be greater than 0" ) # the counter for the chains with the exact desired length lowercase__ = 0 # the cached sizes of the previous chains lowercase__ = {} for start_chain_element in range(1 , __lowerCAmelCase ): # The temporary set will contain the elements of the chain lowercase__ = set() lowercase__ = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. lowercase__ = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(__lowerCAmelCase ) chain_set_length += 1 lowercase__ = digit_factorial_sum(__lowerCAmelCase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] lowercase__ = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F'{solution()}')
709
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _snake_case ( lowercase__): def __init__( self : Optional[Any], __lowercase : str = "▁", __lowercase : bool = True, __lowercase : Union[str, AddedToken] = "<unk>", __lowercase : Union[str, AddedToken] = "</s>", __lowercase : Union[str, AddedToken] = "<pad>", ): lowercase__ = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } lowercase__ = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowercase__ = token_dict["token"] lowercase__ = Tokenizer(Unigram() ) lowercase__ = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ), " " ), normalizers.Lowercase(), ] ) lowercase__ = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) lowercase__ = decoders.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ) lowercase__ = TemplateProcessing( single=F'''$A {self.special_tokens["eos"]["token"]}''', special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])], ) lowercase__ = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase, __lowercase ) def A__ ( self : Union[str, Any], __lowercase : Union[str, List[str]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) if isinstance(__lowercase, __lowercase ): lowercase__ = [files] self._tokenizer.train(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : List[Any], __lowercase : Union[Iterator[str], Iterator[Iterator[str]]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) self._tokenizer.train_from_iterator(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : str ): lowercase__ = json.loads(self._tokenizer.to_str() ) lowercase__ = self.special_tokens["unk"]["id"] lowercase__ = Tokenizer.from_str(json.dumps(__lowercase ) )
37
0
def __lowerCAmelCase ( ): lowercase__ = 0 for i in range(1 , 1001 ): total += i**i return str(lowerCamelCase__ )[-10:] if __name__ == "__main__": print(solution())
710
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase__ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowercase__ = f'''{src_lang}-{tgt_lang}''' lowercase__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) lowercase__ = os.path.join(SCREAMING_SNAKE_CASE_ , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
37
0
import requests def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = {"Content-Type": "application/json"} lowercase__ = requests.post(_lowerCAmelCase , json={"text": message_body} , headers=_lowerCAmelCase ) if response.status_code != 200: lowercase__ = ( "Request to slack returned an error " f'''{response.status_code}, the response is:\n{response.text}''' ) raise ValueError(_lowerCAmelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message("""<YOUR MESSAGE BODY>""", """<SLACK CHANNEL URL>""")
711
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =TransfoXLTokenizer UpperCamelCase__ : List[Any] =False UpperCamelCase__ : List[Any] =False def A__ ( self : Union[str, Any] ): super().setUp() lowercase__ = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : Union[str, Any], **__lowercase : Any ): lowercase__ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname, **__lowercase ) def A__ ( self : Tuple, __lowercase : Optional[int] ): lowercase__ = "<unk> UNwanted , running" lowercase__ = "<unk> unwanted, running" return input_text, output_text def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(vocab_file=self.vocab_file, lower_case=__lowercase ) lowercase__ = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowercase, ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ), [0, 4, 8, 7] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["hello", "!", "how", "are", "you", "?"] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) lowercase__ = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" lowercase__ = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowercase ), __lowercase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowercase ), __lowercase ) def A__ ( self : List[str] ): lowercase__ = self.get_tokenizer() lowercase__ = len(__lowercase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1", 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowercase ), original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ), [1] ) self.assertEqual(tokenizer.decode([1] ), "new1" )
37
0
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowercase_ = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize lowercase_ = "\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n" lowercase_ = "\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n" lowercase_ = "\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n 'meteor': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric('meteor')\n >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\n >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results[\"meteor\"], 4))\n 0.6944\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _snake_case ( datasets.Metric): def A__ ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "predictions": datasets.Value("string", id="sequence" ), "references": datasets.Value("string", id="sequence" ), } ), codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"], reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ], ) def A__ ( self : Any, __lowercase : List[Any] ): import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def A__ ( self : Any, __lowercase : int, __lowercase : List[str], __lowercase : Union[str, Any]=0.9, __lowercase : Any=3, __lowercase : Dict=0.5 ): if NLTK_VERSION >= version.Version("3.6.5" ): lowercase__ = [ meteor_score.single_meteor_score( word_tokenize(_a ), word_tokenize(_a ), alpha=_a, beta=_a, gamma=_a ) for ref, pred in zip(_a, _a ) ] else: lowercase__ = [ meteor_score.single_meteor_score(_a, _a, alpha=_a, beta=_a, gamma=_a ) for ref, pred in zip(_a, _a ) ] return {"meteor": np.mean(_a )}
712
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser(SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.parse_args_into_dataclasses()[0] lowercase__ = TensorFlowBenchmark(args=SCREAMING_SNAKE_CASE_ ) try: lowercase__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowercase__ = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowercase__ = " ".join(str(SCREAMING_SNAKE_CASE_ ).split(" " )[:-1] ) lowercase__ = "" lowercase__ = eval(str(SCREAMING_SNAKE_CASE_ ).split(" " )[-1] ) lowercase__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = full_error_msg + begin_error_msg + str(SCREAMING_SNAKE_CASE_ ) raise ValueError(SCREAMING_SNAKE_CASE_ ) benchmark.run() if __name__ == "__main__": main()
37
0
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.17.0.dev0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/text-classification/requirements.txt""") lowercase_ = logging.getLogger(__name__) @dataclass class _snake_case : UpperCamelCase__ : Optional[str] =field( default="""tab_fact""" , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""}) UpperCamelCase__ : Optional[str] =field( default="""tab_fact""" , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} , ) UpperCamelCase__ : int =field( default=1_0_2_4 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) UpperCamelCase__ : bool =field( default=_snake_case , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""}) UpperCamelCase__ : bool =field( default=_snake_case , metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) } , ) UpperCamelCase__ : Optional[int] =field( default=_snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) UpperCamelCase__ : Optional[int] =field( default=_snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) UpperCamelCase__ : Optional[int] =field( default=_snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) } , ) UpperCamelCase__ : Optional[str] =field( default=_snake_case , metadata={"""help""": """A csv or a json file containing the training data."""}) UpperCamelCase__ : Optional[str] =field( default=_snake_case , metadata={"""help""": """A csv or a json file containing the validation data."""}) UpperCamelCase__ : Optional[str] =field(default=_snake_case , metadata={"""help""": """A csv or a json file containing the test data."""}) def A__ ( self : int ): if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError("Need either a GLUE task, a training/validation file or a dataset name." ) else: lowercase__ = self.train_file.split("." )[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." lowercase__ = self.validation_file.split("." )[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class _snake_case : UpperCamelCase__ : str =field( default=_snake_case , metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""}) UpperCamelCase__ : Optional[str] =field( default=_snake_case , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""}) UpperCamelCase__ : Optional[str] =field( default=_snake_case , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""}) UpperCamelCase__ : Optional[str] =field( default=_snake_case , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) UpperCamelCase__ : bool =field( default=_snake_case , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) UpperCamelCase__ : str =field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) UpperCamelCase__ : bool =field( default=_snake_case , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowercase__ , lowercase__ , lowercase__ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowercase__ , lowercase__ , lowercase__ = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) lowercase__ = training_args.get_process_log_level() logger.setLevel(__UpperCAmelCase ) datasets.utils.logging.set_verbosity(__UpperCAmelCase ) transformers.utils.logging.set_verbosity(__UpperCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(f'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. lowercase__ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase__ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. lowercase__ = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. lowercase__ = {"train": data_args.train_file, "validation": data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: lowercase__ = data_args.train_file.split("." )[-1] lowercase__ = data_args.test_file.split("." )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." lowercase__ = data_args.test_file else: raise ValueError("Need either a GLUE task or a test file for `do_predict`." ) for key in data_files.keys(): logger.info(f'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith(".csv" ): # Loading a dataset from local csv files lowercase__ = load_dataset("csv" , data_files=__UpperCAmelCase , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files lowercase__ = load_dataset("json" , data_files=__UpperCAmelCase , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels lowercase__ = raw_datasets["train"].features["label"].names lowercase__ = len(__UpperCAmelCase ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase__ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__UpperCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer lowercase__ = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=__UpperCAmelCase , ) lowercase__ = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__UpperCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: lowercase__ = "max_length" else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch lowercase__ = False # Some models have set the order of the labels to use, so let's make sure we do use it. lowercase__ = {"Refused": 0, "Entailed": 1} lowercase__ = {0: "Refused", 1: "Entailed"} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' f'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) lowercase__ = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(SCREAMING_SNAKE_CASE_ ): # Tokenize the texts def _convert_table_text_to_pandas(SCREAMING_SNAKE_CASE_ ): lowercase__ = [_table_row.split("#" ) for _table_row in _table_text.strip("\n" ).split("\n" )] lowercase__ = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd lowercase__ = examples["statement"] lowercase__ = list(map(_convert_table_text_to_pandas , examples["table_text"] ) ) lowercase__ = tokenizer(__UpperCAmelCase , __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase ) lowercase__ = examples["label"] return result with training_args.main_process_first(desc="dataset map pre-processing" ): lowercase__ = raw_datasets.map( __UpperCAmelCase , batched=__UpperCAmelCase , load_from_cache_file=not data_args.overwrite_cache , desc="Running tokenizer on dataset" , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError("--do_train requires a train dataset" ) lowercase__ = raw_datasets["train"] if data_args.max_train_samples is not None: lowercase__ = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError("--do_eval requires a validation dataset" ) lowercase__ = raw_datasets["validation"] if data_args.max_eval_samples is not None: lowercase__ = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError("--do_predict requires a test dataset" ) lowercase__ = raw_datasets["test"] if data_args.max_predict_samples is not None: lowercase__ = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(__UpperCAmelCase ) ) , 3 ): logger.info(f'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(SCREAMING_SNAKE_CASE_ ): lowercase__ = p.predictions[0] if isinstance(p.predictions , __UpperCAmelCase ) else p.predictions lowercase__ = np.argmax(__UpperCAmelCase , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: lowercase__ = default_data_collator elif training_args.fpaa: lowercase__ = DataCollatorWithPadding(__UpperCAmelCase , pad_to_multiple_of=8 ) else: lowercase__ = None # Initialize our Trainer lowercase__ = Trainer( model=__UpperCAmelCase , args=__UpperCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=__UpperCAmelCase , tokenizer=__UpperCAmelCase , data_collator=__UpperCAmelCase , ) # Training if training_args.do_train: lowercase__ = None if training_args.resume_from_checkpoint is not None: lowercase__ = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowercase__ = last_checkpoint lowercase__ = trainer.train(resume_from_checkpoint=__UpperCAmelCase ) lowercase__ = train_result.metrics lowercase__ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(__UpperCAmelCase ) ) lowercase__ = min(__UpperCAmelCase , len(__UpperCAmelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics("train" , __UpperCAmelCase ) trainer.save_metrics("train" , __UpperCAmelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("*** Evaluate ***" ) lowercase__ = trainer.evaluate(eval_dataset=__UpperCAmelCase ) lowercase__ = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(__UpperCAmelCase ) lowercase__ = min(__UpperCAmelCase , len(__UpperCAmelCase ) ) trainer.log_metrics("eval" , __UpperCAmelCase ) trainer.save_metrics("eval" , __UpperCAmelCase ) if training_args.do_predict: logger.info("*** Predict ***" ) # Removing the `label` columns because it contains -1 and Trainer won't like that. lowercase__ = predict_dataset.remove_columns("label" ) lowercase__ = trainer.predict(__UpperCAmelCase , metric_key_prefix="predict" ).predictions lowercase__ = np.argmax(__UpperCAmelCase , axis=1 ) lowercase__ = os.path.join(training_args.output_dir , "predict_results_tabfact.txt" ) if trainer.is_world_process_zero(): with open(__UpperCAmelCase , "w" ) as writer: logger.info("***** Predict Results *****" ) writer.write("index\tprediction\n" ) for index, item in enumerate(__UpperCAmelCase ): lowercase__ = label_list[item] writer.write(f'''{index}\t{item}\n''' ) lowercase__ = {"finetuned_from": model_args.model_name_or_path, "tasks": "text-classification"} if training_args.push_to_hub: trainer.push_to_hub(**__UpperCAmelCase ) else: trainer.create_model_card(**__UpperCAmelCase ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): main() if __name__ == "__main__": main()
713
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowercase_ = """<<<<<<< This should probably be modified because it mentions: """ lowercase_ = """======= >>>>>>> """ lowercase_ = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] lowercase_ = [ # (pattern, replacement) # Order is important here for some replacements (r"""tfds\.core""", r"""datasets"""), (r"""tf\.io\.gfile\.GFile""", r"""open"""), (r"""tf\.([\w\d]+)""", r"""datasets.Value('\1')"""), (r"""tfds\.features\.Text\(\)""", r"""datasets.Value('string')"""), (r"""tfds\.features\.Text\(""", r"""datasets.Value('string'),"""), (r"""features\s*=\s*tfds.features.FeaturesDict\(""", r"""features=datasets.Features("""), (r"""tfds\.features\.FeaturesDict\(""", r"""dict("""), (r"""The TensorFlow Datasets Authors""", r"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (r"""tfds\.""", r"""datasets."""), (r"""dl_manager\.manual_dir""", r"""self.config.data_dir"""), (r"""self\.builder_config""", r"""self.config"""), ] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return ConvertCommand(args.tfds_path , args.datasets_directory ) class _snake_case ( lowercase__): @staticmethod def A__ ( __lowercase : ArgumentParser ): lowercase__ = parser.add_parser( "convert", help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.", ) train_parser.add_argument( "--tfds_path", type=__lowercase, required=__lowercase, help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.", ) train_parser.add_argument( "--datasets_directory", type=__lowercase, required=__lowercase, help="Path to the HuggingFace Datasets folder." ) train_parser.set_defaults(func=__lowercase ) def __init__( self : Tuple, __lowercase : str, __lowercase : str, *__lowercase : Tuple ): lowercase__ = get_logger("datasets-cli/converting" ) lowercase__ = tfds_path lowercase__ = datasets_directory def A__ ( self : Any ): if os.path.isdir(self._tfds_path ): lowercase__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__ = os.path.dirname(self._tfds_path ) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path." ) lowercase__ = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) lowercase__ = [] lowercase__ = [] lowercase__ = {} if os.path.isdir(self._tfds_path ): lowercase__ = os.listdir(__lowercase ) else: lowercase__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) if not os.path.isfile(__lowercase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file" ) continue with open(__lowercase, encoding="utf-8" ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = False lowercase__ = False lowercase__ = [] for line in lines: lowercase__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__ = "import datasets\n" elif "import tensorflow" in out_line: # order is important here lowercase__ = "" continue elif "from absl import logging" in out_line: lowercase__ = "from datasets import logging\n" elif "getLogger" in out_line: lowercase__ = out_line.replace("getLogger", "get_logger" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__ = True lowercase__ = list(filter(lambda __lowercase : e in out_line, __lowercase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowercase ) + "\n" ) out_lines.append(__lowercase ) out_lines.append(__lowercase ) continue else: for pattern, replacement in TO_CONVERT: lowercase__ = re.sub(__lowercase, __lowercase, __lowercase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__ = re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)", __lowercase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split("," ) ) lowercase__ = "from . import " + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__ = True out_lines.append(__lowercase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__ = f_name.replace(".py", "" ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) os.makedirs(__lowercase, exist_ok=__lowercase ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowercase ) if needs_manual_update: with_manual_update.append(__lowercase ) with open(__lowercase, "w", encoding="utf-8" ) as f: f.writelines(__lowercase ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: lowercase__ = os.path.basename(__lowercase ) lowercase__ = imports_to_builder_map[f_name.replace(".py", "" )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowercase, __lowercase ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
37
0
from __future__ import annotations def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [] lowercase__ , lowercase__ = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) lowercase__ = result + left + right return input_list def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if len(lowerCAmelCase__ ) <= 1: return input_list lowercase__ = list(lowerCAmelCase__ ) # iteration for two-way merging lowercase__ = 2 while p <= len(lowerCAmelCase__ ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(lowerCAmelCase__ ) , lowerCAmelCase__ ): lowercase__ = i lowercase__ = i + p - 1 lowercase__ = (low + high + 1) // 2 lowercase__ = merge(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # final merge of last two parts if p * 2 >= len(lowerCAmelCase__ ): lowercase__ = i lowercase__ = merge(lowerCAmelCase__ , 0 , lowerCAmelCase__ , len(lowerCAmelCase__ ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() if user_input == "": lowercase_ = [] else: lowercase_ = [int(item.strip()) for item in user_input.split(""",""")] print(iter_merge_sort(unsorted))
714
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase_ = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } lowercase_ = { """allenai/led-base-16384""": 1_6384, } class _snake_case ( lowercase__): UpperCamelCase__ : int =VOCAB_FILES_NAMES UpperCamelCase__ : Any =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =LEDTokenizer UpperCamelCase__ : Tuple =["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any], __lowercase : Optional[Any]=None, __lowercase : Dict=None, __lowercase : Tuple=None, __lowercase : Union[str, Any]="replace", __lowercase : Tuple="<s>", __lowercase : Optional[Any]="</s>", __lowercase : Tuple="</s>", __lowercase : List[str]="<s>", __lowercase : Tuple="<unk>", __lowercase : Dict="<pad>", __lowercase : Dict="<mask>", __lowercase : Any=False, __lowercase : Any=True, **__lowercase : List[Any], ): super().__init__( __lowercase, __lowercase, tokenizer_file=__lowercase, errors=__lowercase, bos_token=__lowercase, eos_token=__lowercase, sep_token=__lowercase, cls_token=__lowercase, unk_token=__lowercase, pad_token=__lowercase, mask_token=__lowercase, add_prefix_space=__lowercase, trim_offsets=__lowercase, **__lowercase, ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = getattr(__lowercase, pre_tok_state.pop("type" ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**__lowercase ) lowercase__ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase__ = "post_processor" lowercase__ = getattr(self.backend_tokenizer, __lowercase, __lowercase ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state["sep"] ) if "cls" in state: lowercase__ = tuple(state["cls"] ) lowercase__ = False if state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get("trim_offsets", __lowercase ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(__lowercase, state.pop("type" ) ) lowercase__ = component_class(**__lowercase ) setattr(self.backend_tokenizer, __lowercase, __lowercase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A__ ( self : str ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self : Optional[int], __lowercase : Dict ): lowercase__ = AddedToken(__lowercase, lstrip=__lowercase, rstrip=__lowercase ) if isinstance(__lowercase, __lowercase ) else value lowercase__ = value def A__ ( self : Any, *__lowercase : List[Any], **__lowercase : Optional[Any] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowercase, **__lowercase ) def A__ ( self : int, *__lowercase : Union[str, Any], **__lowercase : List[str] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowercase, **__lowercase ) def A__ ( self : Optional[Any], __lowercase : str, __lowercase : Optional[str] = None ): lowercase__ = self._tokenizer.model.save(__lowercase, name=__lowercase ) return tuple(__lowercase ) def A__ ( self : List[str], __lowercase : int, __lowercase : Optional[int]=None ): lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self : int, __lowercase : List[int], __lowercase : Optional[List[int]] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self : Union[str, Any], __lowercase : Union[Dict[str, EncodedInput], BatchEncoding], __lowercase : Optional[int] = None, __lowercase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD, __lowercase : Optional[int] = None, __lowercase : Optional[bool] = None, ): lowercase__ = super()._pad( encoded_inputs=__lowercase, max_length=__lowercase, padding_strategy=__lowercase, pad_to_multiple_of=__lowercase, return_attention_mask=__lowercase, ) # Load from model defaults if return_attention_mask is None: lowercase__ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ = len(encoded_inputs["global_attention_mask"] ) != len(__lowercase ) if needs_to_be_padded: lowercase__ = len(__lowercase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
37
0
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = StableDiffusionPipeline.from_pretrained(__A , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors lowercase__ = load_file(__A ) lowercase__ = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: lowercase__ = key.split("." )[0].split(LORA_PREFIX_TEXT_ENCODER + "_" )[-1].split("_" ) lowercase__ = pipeline.text_encoder else: lowercase__ = key.split("." )[0].split(LORA_PREFIX_UNET + "_" )[-1].split("_" ) lowercase__ = pipeline.unet # find the target layer lowercase__ = layer_infos.pop(0 ) while len(__A ) > -1: try: lowercase__ = curr_layer.__getattr__(__A ) if len(__A ) > 0: lowercase__ = layer_infos.pop(0 ) elif len(__A ) == 0: break except Exception: if len(__A ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: lowercase__ = layer_infos.pop(0 ) lowercase__ = [] if "lora_down" in key: pair_keys.append(key.replace("lora_down" , "lora_up" ) ) pair_keys.append(__A ) else: pair_keys.append(__A ) pair_keys.append(key.replace("lora_up" , "lora_down" ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: lowercase__ = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) lowercase__ = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(__A , __A ).unsqueeze(2 ).unsqueeze(3 ) else: lowercase__ = state_dict[pair_keys[0]].to(torch.floataa ) lowercase__ = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(__A , __A ) # update visited list for item in pair_keys: visited.append(__A ) return pipeline if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() parser.add_argument( """--base_model_path""", default=None, type=str, required=True, help="""Path to the base model in diffusers format.""" ) parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--lora_prefix_unet""", default="""lora_unet""", type=str, help="""The prefix of UNet weight in safetensors""" ) parser.add_argument( """--lora_prefix_text_encoder""", default="""lora_te""", type=str, help="""The prefix of text encoder weight in safetensors""", ) parser.add_argument("""--alpha""", default=0.75, type=float, help="""The merging ratio in W = W0 + alpha * deltaW""") parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""" ) parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") lowercase_ = parser.parse_args() lowercase_ = args.base_model_path lowercase_ = args.checkpoint_path lowercase_ = args.dump_path lowercase_ = args.lora_prefix_unet lowercase_ = args.lora_prefix_text_encoder lowercase_ = args.alpha lowercase_ = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) lowercase_ = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
715
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __lowerCAmelCase ( ): lowercase__ = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=SCREAMING_SNAKE_CASE_ ) env_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) launch_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) tpu_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) test_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) # Let's go lowercase__ = parser.parse_args() if not hasattr(SCREAMING_SNAKE_CASE_ , "func" ): parser.print_help() exit(1 ) # Run args.func(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
37
0
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowercase_ = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowercase_ = {"""facebook/blenderbot_small-90M""": 512} def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = set() lowercase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ = char lowercase__ = set(_lowerCamelCase ) return pairs class _snake_case ( a__): UpperCamelCase__ : Tuple =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : str =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =["""input_ids""", """attention_mask"""] def __init__( self : Tuple, __lowercase : Dict, __lowercase : int, __lowercase : str="__start__", __lowercase : List[str]="__end__", __lowercase : List[str]="__unk__", __lowercase : Union[str, Any]="__null__", **__lowercase : Optional[int], ): super().__init__(unk_token=_A, bos_token=_A, eos_token=_A, pad_token=_A, **_A ) with open(_A, encoding="utf-8" ) as vocab_handle: lowercase__ = json.load(_A ) lowercase__ = {v: k for k, v in self.encoder.items()} with open(_A, encoding="utf-8" ) as merges_handle: lowercase__ = merges_handle.read().split("\n" )[1:-1] lowercase__ = [tuple(merge.split() ) for merge in merges] lowercase__ = dict(zip(_A, range(len(_A ) ) ) ) lowercase__ = {} @property def A__ ( self : Optional[Any] ): return len(self.encoder ) def A__ ( self : Any ): return dict(self.encoder, **self.added_tokens_encoder ) def A__ ( self : str, __lowercase : int ): if token in self.cache: return self.cache[token] lowercase__ = re.sub("([.,!?()])", R" \1", _A ) lowercase__ = re.sub("(\')", R" \1 ", _A ) lowercase__ = re.sub(R"\s{2,}", " ", _A ) if "\n" in token: lowercase__ = token.replace("\n", " __newln__" ) lowercase__ = token.split(" " ) lowercase__ = [] for token in tokens: if not len(_A ): continue lowercase__ = token.lower() lowercase__ = tuple(_A ) lowercase__ = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) lowercase__ = get_pairs(_A ) if not pairs: words.append(_A ) continue while True: lowercase__ = min(_A, key=lambda __lowercase : self.bpe_ranks.get(_A, float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowercase__ = bigram lowercase__ = [] lowercase__ = 0 while i < len(_A ): try: lowercase__ = word.index(_A, _A ) new_word.extend(word[i:j] ) lowercase__ = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ = tuple(_A ) lowercase__ = new_word if len(_A ) == 1: break else: lowercase__ = get_pairs(_A ) lowercase__ = '@@ '.join(_A ) lowercase__ = word[:-4] lowercase__ = word words.append(_A ) return " ".join(_A ) def A__ ( self : Optional[int], __lowercase : Union[str, Any] ): lowercase__ = [] lowercase__ = re.findall(R"\S+\n?", _A ) for token in words: split_tokens.extend(list(self.bpe(_A ).split(" " ) ) ) return split_tokens def A__ ( self : List[str], __lowercase : Dict ): lowercase__ = token.lower() return self.encoder.get(_A, self.encoder.get(self.unk_token ) ) def A__ ( self : str, __lowercase : int ): return self.decoder.get(_A, self.unk_token ) def A__ ( self : int, __lowercase : Optional[int] ): lowercase__ = ' '.join(_A ).replace("@@ ", "" ).strip() return out_string def A__ ( self : List[Any], __lowercase : Any, __lowercase : Any = None ): if not os.path.isdir(_A ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( _A, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ = os.path.join( _A, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_A, "w", encoding="utf-8" ) as f: f.write(json.dumps(self.encoder, indent=2, sort_keys=_A, ensure_ascii=_A ) + "\n" ) lowercase__ = 0 with open(_A, "w", encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items(), key=lambda __lowercase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowercase__ = token_index writer.write(" ".join(_A ) + "\n" ) index += 1 return vocab_file, merge_file
716
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : Dict, __lowercase : int, __lowercase : Union[str, Any]=7, __lowercase : Union[str, Any]=3, __lowercase : Any=18, __lowercase : Union[str, Any]=30, __lowercase : Any=400, __lowercase : List[str]=True, __lowercase : Dict=None, __lowercase : List[str]=True, __lowercase : int=False, __lowercase : Union[str, Any]=True, __lowercase : str=True, __lowercase : Optional[int]=[0.5, 0.5, 0.5], __lowercase : List[Any]=[0.5, 0.5, 0.5], ): lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size if size is not None else {"height": 18, "width": 20} lowercase__ = do_thumbnail lowercase__ = do_align_axis lowercase__ = do_pad lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std def A__ ( self : Optional[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =DonutImageProcessor if is_vision_available() else None def A__ ( self : str ): lowercase__ = DonutImageProcessingTester(self ) @property def A__ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_resize" ) ) self.assertTrue(hasattr(__lowercase, "size" ) ) self.assertTrue(hasattr(__lowercase, "do_thumbnail" ) ) self.assertTrue(hasattr(__lowercase, "do_align_long_axis" ) ) self.assertTrue(hasattr(__lowercase, "do_pad" ) ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "image_mean" ) ) self.assertTrue(hasattr(__lowercase, "image_std" ) ) def A__ ( self : str ): lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {"height": 18, "width": 20} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {"height": 42, "width": 42} ) # Previous config had dimensions in (width, height) order lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=(42, 84) ) self.assertEqual(image_processor.size, {"height": 84, "width": 42} ) def A__ ( self : List[str] ): pass @is_flaky() def A__ ( self : Dict ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Optional[Any] ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Tuple ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), )
37
0
from __future__ import annotations import pandas as pd def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [0] * no_of_processes lowercase__ = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(_lowerCamelCase ): lowercase__ = burst_time[i] lowercase__ = 0 lowercase__ = 0 lowercase__ = 9_9999_9999 lowercase__ = 0 lowercase__ = False # Process until all processes are completed while complete != no_of_processes: for j in range(_lowerCamelCase ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: lowercase__ = remaining_time[j] lowercase__ = j lowercase__ = True if not check: increment_time += 1 continue remaining_time[short] -= 1 lowercase__ = remaining_time[short] if minm == 0: lowercase__ = 9_9999_9999 if remaining_time[short] == 0: complete += 1 lowercase__ = False # Find finish time of current process lowercase__ = increment_time + 1 # Calculate waiting time lowercase__ = finish_time - arrival_time[short] lowercase__ = finar - burst_time[short] if waiting_time[short] < 0: lowercase__ = 0 # Increment time increment_time += 1 return waiting_time def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [0] * no_of_processes for i in range(_lowerCamelCase ): lowercase__ = burst_time[i] + waiting_time[i] return turn_around_time def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 lowercase__ = 0 for i in range(_lowerCamelCase ): lowercase__ = total_waiting_time + waiting_time[i] lowercase__ = total_turn_around_time + turn_around_time[i] print(f'''Average waiting time = {total_waiting_time / no_of_processes:.5f}''' ) print("Average turn around time =" , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") lowercase_ = int(input()) lowercase_ = [0] * no_of_processes lowercase_ = [0] * no_of_processes lowercase_ = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) lowercase_ , lowercase_ = map(int, input().split()) lowercase_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowercase_ = burst_time lowercase_ = no_of_processes lowercase_ = waiting_time lowercase_ = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) lowercase_ = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
717
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _snake_case ( lowercase__): def A__ ( self : Optional[Any], __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) lowercase__ = input_file.read() lowercase__ = regexp.search(__lowercase ) return match def A__ ( self : str, __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()", re.DOTALL ) lowercase__ = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowercase__ = regexp.finditer(__lowercase ) lowercase__ = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(F'''open(...) must use utf-8 encoding in {dataset}''' ) def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(F'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
37
0
from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "microsoft/xprophetnet-large-wiki100-cased": ( "https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json" ), } class _snake_case ( _UpperCAmelCase): UpperCamelCase__ : Optional[Any] ="""xlm-prophetnet""" UpperCamelCase__ : Optional[Any] =["""past_key_values"""] UpperCamelCase__ : Optional[Any] ={ """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : List[Any], __lowercase : Optional[float] = 0.1, __lowercase : Optional[Union[str, Callable]] = "gelu", __lowercase : Optional[int] = 3_0522, __lowercase : Optional[int] = 1024, __lowercase : Optional[int] = 4096, __lowercase : Optional[int] = 12, __lowercase : Optional[int] = 16, __lowercase : Optional[int] = 4096, __lowercase : Optional[int] = 12, __lowercase : Optional[int] = 16, __lowercase : Optional[float] = 0.1, __lowercase : Optional[float] = 0.1, __lowercase : Optional[int] = 512, __lowercase : Optional[float] = 0.02, __lowercase : Optional[bool] = True, __lowercase : Optional[bool] = True, __lowercase : Optional[int] = 0, __lowercase : Optional[int] = 2, __lowercase : Optional[int] = 32, __lowercase : Optional[int] = 128, __lowercase : Optional[bool] = False, __lowercase : Optional[float] = 0.0, __lowercase : Optional[bool] = True, __lowercase : Optional[int] = 0, __lowercase : Optional[int] = 1, __lowercase : Optional[int] = 2, **__lowercase : List[Any], ): lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = encoder_ffn_dim lowercase__ = num_encoder_layers lowercase__ = num_encoder_attention_heads lowercase__ = decoder_ffn_dim lowercase__ = num_decoder_layers lowercase__ = num_decoder_attention_heads lowercase__ = max_position_embeddings lowercase__ = init_std # Normal(0, this parameter) lowercase__ = activation_function # parameters for xlmprophetnet lowercase__ = ngram lowercase__ = num_buckets lowercase__ = relative_max_distance lowercase__ = disable_ngram_loss lowercase__ = eps # 3 Types of Dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = dropout lowercase__ = use_cache super().__init__( pad_token_id=lowerCamelCase_, bos_token_id=lowerCamelCase_, eos_token_id=lowerCamelCase_, is_encoder_decoder=lowerCamelCase_, add_cross_attention=lowerCamelCase_, decoder_start_token_id=lowerCamelCase_, **lowerCamelCase_, ) @property def A__ ( self : List[Any] ): return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def A__ ( self : Union[str, Any], __lowercase : Optional[Any] ): raise NotImplementedError( "This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and" " `num_decoder_layers`." )
718
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) lowercase_ = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } lowercase_ = { '''b0''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = EfficientNetConfig() lowercase__ = CONFIG_MAP[model_name]["hidden_dim"] lowercase__ = CONFIG_MAP[model_name]["width_coef"] lowercase__ = CONFIG_MAP[model_name]["depth_coef"] lowercase__ = CONFIG_MAP[model_name]["image_size"] lowercase__ = CONFIG_MAP[model_name]["dropout_rate"] lowercase__ = CONFIG_MAP[model_name]["dw_padding"] lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = 1000 lowercase__ = json.load(open(hf_hub_download(_A , _A , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(_A ): v for k, v in idalabel.items()} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} return config def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(_A , stream=_A ).raw ) return im def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = CONFIG_MAP[model_name]["image_size"] lowercase__ = EfficientNetImageProcessor( size={"height": size, "width": size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.4785_3944, 0.473_2864, 0.4743_4163] , do_center_crop=_A , ) return preprocessor def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = [v.split("_" )[0].split("block" )[1] for v in original_param_names if v.startswith("block" )] lowercase__ = sorted(set(_A ) ) lowercase__ = len(_A ) lowercase__ = {b: str(_A ) for b, i in zip(_A , range(_A ) )} lowercase__ = [] rename_keys.append(("stem_conv/kernel:0", "embeddings.convolution.weight") ) rename_keys.append(("stem_bn/gamma:0", "embeddings.batchnorm.weight") ) rename_keys.append(("stem_bn/beta:0", "embeddings.batchnorm.bias") ) rename_keys.append(("stem_bn/moving_mean:0", "embeddings.batchnorm.running_mean") ) rename_keys.append(("stem_bn/moving_variance:0", "embeddings.batchnorm.running_var") ) for b in block_names: lowercase__ = block_name_mapping[b] rename_keys.append((f'''block{b}_expand_conv/kernel:0''', f'''encoder.blocks.{hf_b}.expansion.expand_conv.weight''') ) rename_keys.append((f'''block{b}_expand_bn/gamma:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.weight''') ) rename_keys.append((f'''block{b}_expand_bn/beta:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.bias''') ) rename_keys.append( (f'''block{b}_expand_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.running_mean''') ) rename_keys.append( (f'''block{b}_expand_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.running_var''') ) rename_keys.append( (f'''block{b}_dwconv/depthwise_kernel:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight''') ) rename_keys.append((f'''block{b}_bn/gamma:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight''') ) rename_keys.append((f'''block{b}_bn/beta:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias''') ) rename_keys.append( (f'''block{b}_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean''') ) rename_keys.append( (f'''block{b}_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var''') ) rename_keys.append((f'''block{b}_se_reduce/kernel:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.reduce.weight''') ) rename_keys.append((f'''block{b}_se_reduce/bias:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.reduce.bias''') ) rename_keys.append((f'''block{b}_se_expand/kernel:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.expand.weight''') ) rename_keys.append((f'''block{b}_se_expand/bias:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.expand.bias''') ) rename_keys.append( (f'''block{b}_project_conv/kernel:0''', f'''encoder.blocks.{hf_b}.projection.project_conv.weight''') ) rename_keys.append((f'''block{b}_project_bn/gamma:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.weight''') ) rename_keys.append((f'''block{b}_project_bn/beta:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.bias''') ) rename_keys.append( (f'''block{b}_project_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.running_mean''') ) rename_keys.append( (f'''block{b}_project_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.running_var''') ) rename_keys.append(("top_conv/kernel:0", "encoder.top_conv.weight") ) rename_keys.append(("top_bn/gamma:0", "encoder.top_bn.weight") ) rename_keys.append(("top_bn/beta:0", "encoder.top_bn.bias") ) rename_keys.append(("top_bn/moving_mean:0", "encoder.top_bn.running_mean") ) rename_keys.append(("top_bn/moving_variance:0", "encoder.top_bn.running_var") ) lowercase__ = {} for item in rename_keys: if item[0] in original_param_names: lowercase__ = "efficientnet." + item[1] lowercase__ = "classifier.weight" lowercase__ = "classifier.bias" return key_mapping def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): for key, value in tf_params.items(): if "normalization" in key: continue lowercase__ = key_mapping[key] if "_conv" in key and "kernel" in key: lowercase__ = torch.from_numpy(_A ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: lowercase__ = torch.from_numpy(_A ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: lowercase__ = torch.from_numpy(np.transpose(_A ) ) else: lowercase__ = torch.from_numpy(_A ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(_A ) @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = model_classes[model_name]( include_top=_A , weights="imagenet" , input_tensor=_A , input_shape=_A , pooling=_A , classes=1000 , classifier_activation="softmax" , ) lowercase__ = original_model.trainable_variables lowercase__ = original_model.non_trainable_variables lowercase__ = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: lowercase__ = param.numpy() lowercase__ = list(tf_params.keys() ) # Load HuggingFace model lowercase__ = get_efficientnet_config(_A ) lowercase__ = EfficientNetForImageClassification(_A ).eval() lowercase__ = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("Converting parameters..." ) lowercase__ = rename_keys(_A ) replace_params(_A , _A , _A ) # Initialize preprocessor and preprocess input image lowercase__ = convert_image_processor(_A ) lowercase__ = preprocessor(images=prepare_img() , return_tensors="pt" ) # HF model inference hf_model.eval() with torch.no_grad(): lowercase__ = hf_model(**_A ) lowercase__ = outputs.logits.detach().numpy() # Original model inference lowercase__ = False lowercase__ = CONFIG_MAP[model_name]["image_size"] lowercase__ = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) lowercase__ = image.img_to_array(_A ) lowercase__ = np.expand_dims(_A , axis=0 ) lowercase__ = original_model.predict(_A ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(_A , _A , atol=1e-3 ), "The predicted logits are not the same." print("Model outputs match!" ) if save_model: # Create folder to save model if not os.path.isdir(_A ): os.mkdir(_A ) # Save converted model and image processor hf_model.save_pretrained(_A ) preprocessor.save_pretrained(_A ) if push_to_hub: # Push model and image processor to hub print(f'''Pushing converted {model_name} to the hub...''' ) lowercase__ = f'''efficientnet-{model_name}''' preprocessor.push_to_hub(_A ) hf_model.push_to_hub(_A ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""b0""", type=str, help="""Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""hf_model""", type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--save_model""", action="""store_true""", help="""Save model to local""") parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") lowercase_ = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
719
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowercase_ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : List[Any], __lowercase : int, __lowercase : Optional[int]=7, __lowercase : List[str]=3, __lowercase : Tuple=18, __lowercase : List[Any]=30, __lowercase : Tuple=400, __lowercase : Any=None, __lowercase : Optional[int]=True, __lowercase : List[str]=True, __lowercase : Union[str, Any]=None, ): lowercase__ = size if size is not None else {"height": 20, "width": 20} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = size lowercase__ = do_normalize lowercase__ = do_convert_rgb lowercase__ = [512, 1024, 2048, 4096] lowercase__ = patch_size if patch_size is not None else {"height": 16, "width": 16} def A__ ( self : List[str] ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def A__ ( self : Any ): lowercase__ = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" lowercase__ = Image.open(requests.get(__lowercase, stream=__lowercase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Any =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Any ): lowercase__ = PixaStructImageProcessingTester(self ) @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Optional[int] ): lowercase__ = self.image_processor_tester.prepare_dummy_image() lowercase__ = self.image_processing_class(**self.image_processor_dict ) lowercase__ = 2048 lowercase__ = image_processor(__lowercase, return_tensors="pt", max_patches=__lowercase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean(), torch.tensor(0.0606 ), atol=1e-3, rtol=1e-3 ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : int ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 lowercase__ = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__lowercase ): lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches lowercase__ = "Hello" lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Tuple ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Any ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Optional[int] ): lowercase__ = PixaStructImageProcessingTester(self, num_channels=4 ) lowercase__ = 3 @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), )
37
0
'''simple docstring''' import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class _snake_case : def __init__( self : Tuple, __lowercase : Optional[Any], __lowercase : Optional[Any]=sys.maxsize ): lowercase__ = "bilinear" lowercase__ = max_size lowercase__ = short_edge_length def __call__( self : str, __lowercase : str ): lowercase__ = [] for img in imgs: lowercase__ , lowercase__ = img.shape[:2] # later: provide list and randomly choose index for resize lowercase__ = np.random.randint(self.short_edge_length[0], self.short_edge_length[1] + 1 ) if size == 0: return img lowercase__ = size * 1.0 / min(snake_case_, snake_case_ ) if h < w: lowercase__ , lowercase__ = size, scale * w else: lowercase__ , lowercase__ = scale * h, size if max(snake_case_, snake_case_ ) > self.max_size: lowercase__ = self.max_size * 1.0 / max(snake_case_, snake_case_ ) lowercase__ = newh * scale lowercase__ = neww * scale lowercase__ = int(neww + 0.5 ) lowercase__ = int(newh + 0.5 ) if img.dtype == np.uinta: lowercase__ = Image.fromarray(snake_case_ ) lowercase__ = pil_image.resize((neww, newh), PILImageResampling.BILINEAR ) lowercase__ = np.asarray(snake_case_ ) else: lowercase__ = img.permute(2, 0, 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw lowercase__ = nn.functional.interpolate( snake_case_, (newh, neww), mode=self.interp_method, align_corners=snake_case_ ).squeeze(0 ) img_augs.append(snake_case_ ) return img_augs class _snake_case : def __init__( self : List[str], __lowercase : List[str] ): lowercase__ = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST], cfg.INPUT.MAX_SIZE_TEST ) lowercase__ = cfg.INPUT.FORMAT lowercase__ = cfg.SIZE_DIVISIBILITY lowercase__ = cfg.PAD_VALUE lowercase__ = cfg.INPUT.MAX_SIZE_TEST lowercase__ = cfg.MODEL.DEVICE lowercase__ = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ), 1, 1 ) lowercase__ = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ), 1, 1 ) lowercase__ = lambda __lowercase : (x - self.pixel_mean) / self.pixel_std def A__ ( self : List[str], __lowercase : int ): lowercase__ = tuple(max(snake_case_ ) for s in zip(*[img.shape for img in images] ) ) lowercase__ = [im.shape[-2:] for im in images] lowercase__ = [ nn.functional.pad( snake_case_, [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]], value=self.pad_value, ) for size, im in zip(snake_case_, snake_case_ ) ] return torch.stack(snake_case_ ), torch.tensor(snake_case_ ) def __call__( self : List[str], __lowercase : Any, __lowercase : Dict=False ): with torch.no_grad(): if not isinstance(snake_case_, snake_case_ ): lowercase__ = [images] if single_image: assert len(snake_case_ ) == 1 for i in range(len(snake_case_ ) ): if isinstance(images[i], torch.Tensor ): images.insert(snake_case_, images.pop(snake_case_ ).to(self.device ).float() ) elif not isinstance(images[i], torch.Tensor ): images.insert( snake_case_, torch.as_tensor(img_tensorize(images.pop(snake_case_ ), input_format=self.input_format ) ) .to(self.device ) .float(), ) # resize smallest edge lowercase__ = torch.tensor([im.shape[:2] for im in images] ) lowercase__ = self.aug(snake_case_ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic lowercase__ = [self.normalizer(snake_case_ ) for x in images] # now pad them to do the following operations lowercase__ , lowercase__ = self.pad(snake_case_ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad lowercase__ = torch.true_divide(snake_case_, snake_case_ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): assert torch.isfinite(SCREAMING_SNAKE_CASE_ ).all(), "Box tensor contains infinite or NaN!" lowercase__ , lowercase__ = box_size tensor[:, 0].clamp_(min=0 , max=SCREAMING_SNAKE_CASE_ ) tensor[:, 1].clamp_(min=0 , max=SCREAMING_SNAKE_CASE_ ) tensor[:, 2].clamp_(min=0 , max=SCREAMING_SNAKE_CASE_ ) tensor[:, 3].clamp_(min=0 , max=SCREAMING_SNAKE_CASE_ )
720
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = len(SCREAMING_SNAKE_CASE_ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ = 0 count += depth_first_search(SCREAMING_SNAKE_CASE_ , row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , row - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col - 1 , SCREAMING_SNAKE_CASE_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
37
0
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =DDIMPipeline UpperCamelCase__ : Dict =UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase__ : int =PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } UpperCamelCase__ : Optional[Any] =UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase__ : Tuple =False def A__ ( self : Optional[int] ): torch.manual_seed(0 ) lowercase__ = UNetaDModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=3, out_channels=3, down_block_types=("DownBlock2D", "AttnDownBlock2D"), up_block_types=("AttnUpBlock2D", "UpBlock2D"), ) lowercase__ = DDIMScheduler() lowercase__ = {"""unet""": unet, """scheduler""": scheduler} return components def A__ ( self : str, __lowercase : Any, __lowercase : List[str]=0 ): if str(_lowercase ).startswith("mps" ): lowercase__ = torch.manual_seed(_lowercase ) else: lowercase__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) lowercase__ = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def A__ ( self : List[str] ): lowercase__ = """cpu""" lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) lowercase__ = self.get_dummy_inputs(_lowercase ) lowercase__ = pipe(**_lowercase ).images lowercase__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape, (1, 32, 32, 3) ) lowercase__ = np.array( [1.0_0_0e0_0, 5.7_1_7e-0_1, 4.7_1_7e-0_1, 1.0_0_0e0_0, 0.0_0_0e0_0, 1.0_0_0e0_0, 3.0_0_0e-0_4, 0.0_0_0e0_0, 9.0_0_0e-0_4] ) lowercase__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowercase, 1e-3 ) def A__ ( self : Optional[int] ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def A__ ( self : Dict ): super().test_save_load_local(expected_max_difference=3e-3 ) def A__ ( self : Union[str, Any] ): super().test_save_load_optional_components(expected_max_difference=3e-3 ) def A__ ( self : Optional[int] ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class _snake_case ( unittest.TestCase): def A__ ( self : Union[str, Any] ): lowercase__ = """google/ddpm-cifar10-32""" lowercase__ = UNetaDModel.from_pretrained(_lowercase ) lowercase__ = DDIMScheduler() lowercase__ = DDIMPipeline(unet=_lowercase, scheduler=_lowercase ) ddim.to(_lowercase ) ddim.set_progress_bar_config(disable=_lowercase ) lowercase__ = torch.manual_seed(0 ) lowercase__ = ddim(generator=_lowercase, eta=0.0, output_type="numpy" ).images lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase__ = np.array([0.1723, 0.1617, 0.1600, 0.1626, 0.1497, 0.1513, 0.1505, 0.1442, 0.1453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self : Tuple ): lowercase__ = """google/ddpm-ema-bedroom-256""" lowercase__ = UNetaDModel.from_pretrained(_lowercase ) lowercase__ = DDIMScheduler.from_pretrained(_lowercase ) lowercase__ = DDIMPipeline(unet=_lowercase, scheduler=_lowercase ) ddpm.to(_lowercase ) ddpm.set_progress_bar_config(disable=_lowercase ) lowercase__ = torch.manual_seed(0 ) lowercase__ = ddpm(generator=_lowercase, output_type="numpy" ).images lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) lowercase__ = np.array([0.0060, 0.0201, 0.0344, 0.0024, 0.0018, 0.0002, 0.0022, 0.0000, 0.0069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
721
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 for ch in input_str: lowercase__ = ord(SCREAMING_SNAKE_CASE_ ) lowercase__ = pow(2 , SCREAMING_SNAKE_CASE_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
37
0
'''simple docstring''' from __future__ import annotations def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = set(snake_case__ ), [start] while stack: lowercase__ = stack.pop() explored.add(snake_case__ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(snake_case__ ) return explored lowercase_ = { """A""": ["""B""", """C""", """D"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F"""], """D""": ["""B""", """D"""], """E""": ["""B""", """F"""], """F""": ["""C""", """E""", """G"""], """G""": ["""F"""], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, """A"""))
700
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = len(SCREAMING_SNAKE_CASE_ ) for i in range(SCREAMING_SNAKE_CASE_ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE_ ): if numbers[j] < numbers[i]: lowercase__ , lowercase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() lowercase_ = [int(item) for item in user_input.split(""",""")] print(exchange_sort(unsorted))
37
0
import requests lowercase_ = """YOUR API KEY""" def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = giphy_api_key ): lowercase__ = "+".join(query.split() ) lowercase__ = f'''https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}''' lowercase__ = requests.get(__SCREAMING_SNAKE_CASE ).json()["data"] return [gif["url"] for gif in gifs] if __name__ == "__main__": print("""\n""".join(get_gifs("""space ship""")))
701
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) lowercase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase__ = 1 if upper_limit > 0: lowercase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: lowercase_ = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F'The Catalan numbers from 0 through {N} are:') print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
37
0
import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = SwinConfig() lowercase__ = swin_name.split("_" ) lowercase__ = name_split[1] lowercase__ = int(name_split[4] ) lowercase__ = int(name_split[3][-1] ) if model_size == "tiny": lowercase__ = 96 lowercase__ = (2, 2, 6, 2) lowercase__ = (3, 6, 12, 24) elif model_size == "small": lowercase__ = 96 lowercase__ = (2, 2, 18, 2) lowercase__ = (3, 6, 12, 24) elif model_size == "base": lowercase__ = 128 lowercase__ = (2, 2, 18, 2) lowercase__ = (4, 8, 16, 32) else: lowercase__ = 192 lowercase__ = (2, 2, 18, 2) lowercase__ = (6, 12, 24, 48) if "in22k" in swin_name: lowercase__ = 2_1841 else: lowercase__ = 1000 lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = idalabel lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = img_size lowercase__ = num_classes lowercase__ = embed_dim lowercase__ = depths lowercase__ = num_heads lowercase__ = window_size return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "patch_embed.proj" in name: lowercase__ = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: lowercase__ = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: lowercase__ = "encoder." + name if "attn.proj" in name: lowercase__ = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: lowercase__ = name.replace("attn" , "attention.self" ) if "norm1" in name: lowercase__ = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: lowercase__ = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: lowercase__ = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: lowercase__ = name.replace("mlp.fc2" , "output.dense" ) if name == "norm.weight": lowercase__ = "layernorm.weight" if name == "norm.bias": lowercase__ = "layernorm.bias" if "head" in name: lowercase__ = name.replace("head" , "classifier" ) else: lowercase__ = "swin." + name return name def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): for key in orig_state_dict.copy().keys(): lowercase__ = orig_state_dict.pop(SCREAMING_SNAKE_CASE_ ) if "mask" in key: continue elif "qkv" in key: lowercase__ = key.split("." ) lowercase__ = int(key_split[1] ) lowercase__ = int(key_split[3] ) lowercase__ = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowercase__ = val[:dim, :] lowercase__ = val[ dim : dim * 2, : ] lowercase__ = val[-dim:, :] else: lowercase__ = val[ :dim ] lowercase__ = val[ dim : dim * 2 ] lowercase__ = val[ -dim: ] else: lowercase__ = val return orig_state_dict def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = timm.create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() lowercase__ = get_swin_config(SCREAMING_SNAKE_CASE_ ) lowercase__ = SwinForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() lowercase__ = convert_state_dict(timm_model.state_dict() , SCREAMING_SNAKE_CASE_ ) model.load_state_dict(SCREAMING_SNAKE_CASE_ ) lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = AutoImageProcessor.from_pretrained("microsoft/{}".format(swin_name.replace("_" , "-" ) ) ) lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) lowercase__ = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors="pt" ) lowercase__ = timm_model(inputs["pixel_values"] ) lowercase__ = model(**SCREAMING_SNAKE_CASE_ ).logits assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ) print(f'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swin_name""", default="""swin_tiny_patch4_window7_224""", type=str, help="""Name of the Swin timm model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowercase_ = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
702
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase_ = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _snake_case ( unittest.TestCase): def A__ ( self : Tuple ): lowercase__ = 10 def A__ ( self : int ): lowercase__ = [1, 2, 3, 4] lowercase__ = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(__A, self.block_size, 0 ), __A ) def A__ ( self : Dict ): lowercase__ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] lowercase__ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(__A, self.block_size, 0 ), __A ) def A__ ( self : Optional[Any] ): lowercase__ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] lowercase__ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(__A, self.block_size, 0 ), __A ) def A__ ( self : Union[str, Any] ): lowercase__ = "It was the year of Our Lord one thousand seven hundred and\n seventy-five.\n\nSpiritual revelations were conceded to England at that\n favoured period, as at this." lowercase__ = process_story(__A ) self.assertEqual(__A, [] ) def A__ ( self : List[str] ): lowercase__ = "" lowercase__ = process_story(__A ) self.assertEqual(__A, [] ) self.assertEqual(__A, [] ) def A__ ( self : Dict ): lowercase__ = ( "It was the year of Our Lord one thousand seven hundred and " "seventy-five\n\nSpiritual revelations were conceded to England " "at that favoured period, as at this.\n@highlight\n\nIt was the best of times" ) lowercase__ = process_story(__A ) lowercase__ = [ "It was the year of Our Lord one thousand seven hundred and seventy-five.", "Spiritual revelations were conceded to England at that favoured period, as at this.", ] self.assertEqual(__A, __A ) lowercase__ = ["It was the best of times."] self.assertEqual(__A, __A ) def A__ ( self : Optional[Any] ): lowercase__ = torch.tensor([1, 2, 3, 4] ) lowercase__ = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(__A, 0 ).numpy(), expected.numpy() ) def A__ ( self : List[str] ): lowercase__ = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) lowercase__ = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(__A, 23 ).numpy(), expected.numpy() ) def A__ ( self : List[Any] ): lowercase__ = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) lowercase__ = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(__A, 1 ).numpy(), expected.numpy() ) def A__ ( self : Optional[Any] ): lowercase__ = 101 lowercase__ = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]] ) lowercase__ = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) lowercase__ = compute_token_type_ids(__A, __A ) np.testing.assert_array_equal(__A, __A )
703
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") lowercase__ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f'''bert/{name}''' def create_tf_var(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = session.run(SCREAMING_SNAKE_CASE_ ) print(f'''Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace("-" , "_" ) + ".ckpt" ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_=None ): lowercase__ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory in which to save tensorflow model" ) lowercase__ = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
37
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging lowercase_ = logging.get_logger(__name__) class _snake_case ( lowercase__): UpperCamelCase__ : Optional[int] =["""pixel_values"""] def __init__( self : int, __lowercase : bool = True, __lowercase : Optional[Dict[str, int]] = None, __lowercase : PILImageResampling = PILImageResampling.BILINEAR, __lowercase : bool = True, __lowercase : Dict[str, int] = None, __lowercase : bool = True, __lowercase : Union[int, float] = 1 / 255, __lowercase : bool = True, __lowercase : Optional[Union[float, List[float]]] = None, __lowercase : Optional[Union[float, List[float]]] = None, **__lowercase : List[str], ): super().__init__(**__lowercase ) lowercase__ = size if size is not None else {"shortest_edge": 256} lowercase__ = get_size_dict(__lowercase, default_to_square=__lowercase ) lowercase__ = crop_size if crop_size is not None else {"height": 224, "width": 224} lowercase__ = get_size_dict(__lowercase ) lowercase__ = do_resize lowercase__ = size lowercase__ = resample lowercase__ = do_center_crop lowercase__ = crop_size lowercase__ = do_rescale lowercase__ = rescale_factor lowercase__ = do_normalize lowercase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowercase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self : Tuple, __lowercase : np.ndarray, __lowercase : Dict[str, int], __lowercase : PILImageResampling = PILImageResampling.BICUBIC, __lowercase : Optional[Union[str, ChannelDimension]] = None, **__lowercase : Optional[int], ): lowercase__ = get_size_dict(__lowercase, default_to_square=__lowercase ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) lowercase__ = get_resize_output_image_size(__lowercase, size=size["shortest_edge"], default_to_square=__lowercase ) return resize(__lowercase, size=__lowercase, resample=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : List[Any], __lowercase : np.ndarray, __lowercase : Dict[str, int], __lowercase : Optional[Union[str, ChannelDimension]] = None, **__lowercase : List[Any], ): lowercase__ = get_size_dict(__lowercase ) return center_crop(__lowercase, size=(size["height"], size["width"]), data_format=__lowercase, **__lowercase ) def A__ ( self : Optional[int], __lowercase : np.ndarray, __lowercase : float, __lowercase : Optional[Union[str, ChannelDimension]] = None, **__lowercase : Dict ): return rescale(__lowercase, scale=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : Union[str, Any], __lowercase : np.ndarray, __lowercase : Union[float, List[float]], __lowercase : Union[float, List[float]], __lowercase : Optional[Union[str, ChannelDimension]] = None, **__lowercase : Union[str, Any], ): return normalize(__lowercase, mean=__lowercase, std=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : Tuple, __lowercase : ImageInput, __lowercase : Optional[bool] = None, __lowercase : Dict[str, int] = None, __lowercase : PILImageResampling = None, __lowercase : bool = None, __lowercase : Dict[str, int] = None, __lowercase : Optional[bool] = None, __lowercase : Optional[float] = None, __lowercase : Optional[bool] = None, __lowercase : Optional[Union[float, List[float]]] = None, __lowercase : Optional[Union[float, List[float]]] = None, __lowercase : Optional[Union[str, TensorType]] = None, __lowercase : Union[str, ChannelDimension] = ChannelDimension.FIRST, **__lowercase : Any, ): lowercase__ = do_resize if do_resize is not None else self.do_resize lowercase__ = size if size is not None else self.size lowercase__ = get_size_dict(__lowercase, default_to_square=__lowercase ) lowercase__ = resample if resample is not None else self.resample lowercase__ = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase__ = crop_size if crop_size is not None else self.crop_size lowercase__ = get_size_dict(__lowercase ) lowercase__ = do_rescale if do_rescale is not None else self.do_rescale lowercase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ = do_normalize if do_normalize is not None else self.do_normalize lowercase__ = image_mean if image_mean is not None else self.image_mean lowercase__ = image_std if image_std is not None else self.image_std lowercase__ = make_list_of_images(__lowercase ) if not valid_images(__lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. lowercase__ = [to_numpy_array(__lowercase ) for image in images] if do_resize: lowercase__ = [self.resize(image=__lowercase, size=__lowercase, resample=__lowercase ) for image in images] if do_center_crop: lowercase__ = [self.center_crop(image=__lowercase, size=__lowercase ) for image in images] if do_rescale: lowercase__ = [self.rescale(image=__lowercase, scale=__lowercase ) for image in images] if do_normalize: lowercase__ = [self.normalize(image=__lowercase, mean=__lowercase, std=__lowercase ) for image in images] lowercase__ = [to_channel_dimension_format(__lowercase, __lowercase ) for image in images] lowercase__ = {"pixel_values": images} return BatchFeature(data=__lowercase, tensor_type=__lowercase )
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_timesformer""": ["""TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimesformerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimesformerModel""", """TimesformerForVideoClassification""", """TimesformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import fire from utils import calculate_rouge, save_json def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ ): lowercase__ = [x.strip() for x in open(SCREAMING_SNAKE_CASE_ ).readlines()] lowercase__ = [x.strip() for x in open(SCREAMING_SNAKE_CASE_ ).readlines()][: len(SCREAMING_SNAKE_CASE_ )] lowercase__ = calculate_rouge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) if save_path is not None: save_json(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , indent=SCREAMING_SNAKE_CASE_ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
705
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() lowercase_ = { """bart""": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), """bert""": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-base-cased-finetuned-mrpc""": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """dpr""": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), """gpt2""": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlnet""": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm""": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm-roberta""": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """transfo-xl""": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """openai-gpt""": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """roberta""": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """layoutlm""": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), """roberta-large-mnli""": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """camembert""": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """flaubert""": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert""": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert-base-distilled-squad""": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert""": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert-visual-feature-encoder""": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """ctrl""": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """albert""": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """t5""": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """electra""": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """wav2vec2""": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) lowercase__ = config_class.from_json_file(SCREAMING_SNAKE_CASE_ ) lowercase__ = True lowercase__ = True print(f'''Building TensorFlow model from configuration: {config}''' ) lowercase__ = model_class(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): lowercase__ = cached_file( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: lowercase__ = load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if compare_with_pt_model: lowercase__ = tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE_ ) # build the network lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" ) lowercase__ = pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE_ , config=SCREAMING_SNAKE_CASE_ , state_dict=SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): lowercase__ = pt_model(**pt_model.dummy_inputs ) lowercase__ = pto[0].numpy() lowercase__ = tfo[0].numpy() lowercase__ = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(SCREAMING_SNAKE_CASE_ , save_format="h5" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ): if args_model_type is None: lowercase__ = list(MODEL_CLASSES.keys() ) else: lowercase__ = [args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE_ , start=1 ): print("=" * 100 ) print(f''' Converting model type {j}/{len(SCREAMING_SNAKE_CASE_ )}: {model_type}''' ) print("=" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: lowercase__ = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: lowercase__ = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , start=1 ): print("-" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue lowercase__ = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE_ )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 100 ) if config_shortcut_name in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = config_shortcut_name if model_shortcut_name in aws_model_maps: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE_ ): lowercase__ = "converted_model" convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE_ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE_ , config_file=SCREAMING_SNAKE_CASE_ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE_ , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=SCREAMING_SNAKE_CASE_ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE_ ) os.remove(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_dump_path""", default=None, type=str, required=True, help="""Path to the output Tensorflow dump file.""" ) parser.add_argument( """--model_type""", default=None, type=str, help=( F'Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ' """convert all the models from AWS.""" ), ) parser.add_argument( """--pytorch_checkpoint_path""", default=None, type=str, help=( """Path to the PyTorch checkpoint path or shortcut name to download from AWS. """ """If not given, will download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--config_file""", default=None, type=str, help=( """The config json file corresponding to the pre-trained model. \n""" """This specifies the model architecture. If not given and """ """--pytorch_checkpoint_path is not given or is a shortcut name """ """use the configuration associated to the shortcut name on the AWS""" ), ) parser.add_argument( """--compare_with_pt_model""", action="""store_true""", help="""Compare Tensorflow and PyTorch model predictions.""" ) parser.add_argument( """--use_cached_models""", action="""store_true""", help="""Use cached models if possible instead of updating to latest checkpoint versions.""", ) parser.add_argument( """--remove_cached_files""", action="""store_true""", help="""Remove pytorch models after conversion (save memory when converting in batches).""", ) parser.add_argument("""--only_convert_finetuned_models""", action="""store_true""", help="""Only convert finetuned models.""") lowercase_ = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
37
0
import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class _snake_case ( __a): def A__ ( self : Optional[Any] ): lowercase__ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(a_, "hidden_sizes" ) ) self.parent.assertTrue(hasattr(a_, "num_attention_heads" ) ) class _snake_case : def __init__( self : int, __lowercase : Optional[Any], __lowercase : Union[str, Any]=13, __lowercase : Dict=64, __lowercase : Union[str, Any]=3, __lowercase : Union[str, Any]=3, __lowercase : Optional[int]=2, __lowercase : List[str]=1, __lowercase : Optional[Any]=16, __lowercase : str=[128, 256, 384], __lowercase : Optional[int]=[4, 6, 8], __lowercase : Optional[Any]=[2, 3, 4], __lowercase : Optional[Any]=[16, 16, 16], __lowercase : Tuple=0, __lowercase : Dict=[2, 2, 2], __lowercase : Dict=[2, 2, 2], __lowercase : Optional[int]=0.02, __lowercase : Optional[int]=True, __lowercase : List[Any]=True, __lowercase : Optional[int]=2, ): lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = num_channels lowercase__ = kernel_size lowercase__ = stride lowercase__ = padding lowercase__ = hidden_sizes lowercase__ = num_attention_heads lowercase__ = depths lowercase__ = key_dim lowercase__ = drop_path_rate lowercase__ = patch_size lowercase__ = attention_ratio lowercase__ = mlp_ratio lowercase__ = initializer_range lowercase__ = [ ["""Subsample""", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["""Subsample""", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] lowercase__ = is_training lowercase__ = use_labels lowercase__ = num_labels lowercase__ = initializer_range def A__ ( self : List[str] ): lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size], self.num_labels ) lowercase__ = self.get_config() return config, pixel_values, labels def A__ ( self : Optional[int] ): return LevitConfig( image_size=self.image_size, num_channels=self.num_channels, kernel_size=self.kernel_size, stride=self.stride, padding=self.padding, patch_size=self.patch_size, hidden_sizes=self.hidden_sizes, num_attention_heads=self.num_attention_heads, depths=self.depths, key_dim=self.key_dim, drop_path_rate=self.drop_path_rate, mlp_ratio=self.mlp_ratio, attention_ratio=self.attention_ratio, initializer_range=self.initializer_range, down_ops=self.down_ops, ) def A__ ( self : str, __lowercase : str, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = LevitModel(config=a_ ) model.to(a_ ) model.eval() lowercase__ = model(a_ ) lowercase__ = (self.image_size, self.image_size) lowercase__ = image_size[0], image_size[1] for _ in range(4 ): lowercase__ = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) lowercase__ = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]), ) def A__ ( self : Union[str, Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : str ): lowercase__ = self.num_labels lowercase__ = LevitForImageClassification(a_ ) model.to(a_ ) model.eval() lowercase__ = model(a_, labels=a_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def A__ ( self : str ): lowercase__ = self.prepare_config_and_inputs() lowercase__ = config_and_inputs lowercase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _snake_case ( __a , __a , unittest.TestCase): UpperCamelCase__ : List[Any] =( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) UpperCamelCase__ : Tuple =( { """feature-extraction""": LevitModel, """image-classification""": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) UpperCamelCase__ : Optional[Any] =False UpperCamelCase__ : Optional[int] =False UpperCamelCase__ : List[Any] =False UpperCamelCase__ : int =False UpperCamelCase__ : int =False def A__ ( self : Tuple ): lowercase__ = LevitModelTester(self ) lowercase__ = ConfigTester(self, config_class=a_, has_text_modality=a_, hidden_size=37 ) def A__ ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self : List[Any] ): return @unittest.skip(reason="Levit does not use inputs_embeds" ) def A__ ( self : Tuple ): pass @unittest.skip(reason="Levit does not support input and output embeddings" ) def A__ ( self : Union[str, Any] ): pass @unittest.skip(reason="Levit does not output attentions" ) def A__ ( self : Optional[int] ): pass def A__ ( self : int ): lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(a_ ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1], a_ ) def A__ ( self : str ): def check_hidden_states_output(__lowercase : Optional[int], __lowercase : Tuple, __lowercase : List[Any] ): lowercase__ = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowercase__ = model(**self._prepare_for_class(a_, a_ ) ) lowercase__ = outputs.hidden_states lowercase__ = len(self.model_tester.depths ) + 1 self.assertEqual(len(a_ ), a_ ) lowercase__ = (self.model_tester.image_size, self.model_tester.image_size) lowercase__ = image_size[0], image_size[1] for _ in range(4 ): lowercase__ = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) lowercase__ = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ), [ height * width, self.model_tester.hidden_sizes[0], ], ) lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = True check_hidden_states_output(a_, a_, a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ = True check_hidden_states_output(a_, a_, a_ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self : Tuple ): pass def A__ ( self : Tuple, __lowercase : List[Any], __lowercase : List[Any], __lowercase : Optional[int]=False ): lowercase__ = super()._prepare_for_class(a_, a_, return_labels=a_ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def A__ ( self : Tuple ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A__ ( self : List[Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) def A__ ( self : Union[str, Any] ): if not self.model_tester.is_training: return lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(a_ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue lowercase__ = model_class(a_ ) model.to(a_ ) model.train() lowercase__ = self._prepare_for_class(a_, a_, return_labels=a_ ) lowercase__ = model(**a_ ).loss loss.backward() def A__ ( self : List[Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowercase__ = False lowercase__ = True for model_class in self.all_model_classes: if model_class in get_values(a_ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue lowercase__ = model_class(a_ ) model.gradient_checkpointing_enable() model.to(a_ ) model.train() lowercase__ = self._prepare_for_class(a_, a_, return_labels=a_ ) lowercase__ = model(**a_ ).loss loss.backward() def A__ ( self : Any ): lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(a_ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type["title"]}''' ): lowercase__ = problem_type["""title"""] lowercase__ = problem_type["""num_labels"""] lowercase__ = model_class(a_ ) model.to(a_ ) model.train() lowercase__ = self._prepare_for_class(a_, a_, return_labels=a_ ) if problem_type["num_labels"] > 1: lowercase__ = inputs["""labels"""].unsqueeze(1 ).repeat(1, problem_type["num_labels"] ) lowercase__ = inputs["""labels"""].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=a_ ) as warning_list: lowercase__ = model(**a_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def A__ ( self : List[str] ): for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = LevitModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def __lowerCAmelCase ( ): lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class _snake_case ( unittest.TestCase): @cached_property def A__ ( self : int ): return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def A__ ( self : Tuple ): lowercase__ = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( a_ ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=a_, return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): lowercase__ = model(**a_ ) # verify the logits lowercase__ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape, a_ ) lowercase__ = torch.tensor([1.0448, -0.3745, -1.8317] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3], a_, atol=1e-4 ) )
706
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : List[Any], __lowercase : str, __lowercase : List[Any]=7, __lowercase : Optional[int]=3, __lowercase : Dict=18, __lowercase : int=30, __lowercase : str=400, __lowercase : Optional[int]=True, __lowercase : str=32, __lowercase : List[str]=True, ): lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size_divisor lowercase__ = do_rescale def A__ ( self : int ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : int =GLPNImageProcessor if is_vision_available() else None def A__ ( self : Optional[int] ): lowercase__ = GLPNImageProcessingTester(self ) @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : int ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCAmelCase, "do_resize" ) ) self.assertTrue(hasattr(__lowerCAmelCase, "size_divisor" ) ) self.assertTrue(hasattr(__lowerCAmelCase, "resample" ) ) self.assertTrue(hasattr(__lowerCAmelCase, "do_rescale" ) ) def A__ ( self : Union[str, Any] ): pass def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase, Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A__ ( self : str ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowerCAmelCase, numpify=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase, np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowerCAmelCase, torchify=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase, torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
707
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : def __init__( self : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Union[str, Any], __lowercase : str, __lowercase : List[Any], __lowercase : List[str]=0.2, __lowercase : List[str]=0.2 ): lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = conva_get[:2] lowercase__ = conva_get[2] lowercase__ = size_pa lowercase__ = rate_w lowercase__ = rate_t lowercase__ = [ np.mat(-1 * np.random.rand(self.conva[0], self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self : Any, __lowercase : List[str] ): # save model dict with pickle lowercase__ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(__lowercase, "wb" ) as f: pickle.dump(__lowercase, __lowercase ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls : Dict, __lowercase : Union[str, Any] ): # read saved model with open(__lowercase, "rb" ) as f: lowercase__ = pickle.load(__lowercase ) # noqa: S301 lowercase__ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) lowercase__ = model_dic.get("size_pooling1" ) lowercase__ = model_dic.get("num_bp1" ) lowercase__ = model_dic.get("num_bp2" ) lowercase__ = model_dic.get("num_bp3" ) lowercase__ = model_dic.get("rate_weight" ) lowercase__ = model_dic.get("rate_thre" ) # create model instance lowercase__ = CNN(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) # modify model parameter lowercase__ = model_dic.get("w_conv1" ) lowercase__ = model_dic.get("wkj" ) lowercase__ = model_dic.get("vji" ) lowercase__ = model_dic.get("thre_conv1" ) lowercase__ = model_dic.get("thre_bp2" ) lowercase__ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self : str, __lowercase : List[Any] ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self : List[str], __lowercase : Optional[Any] ): return round(__lowercase, 3 ) def A__ ( self : Optional[Any], __lowercase : Dict, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : str ): # convolution process lowercase__ = convs[0] lowercase__ = convs[1] lowercase__ = np.shape(__lowercase )[0] # get the data slice of original image data, data_focus lowercase__ = [] for i_focus in range(0, size_data - size_conv + 1, __lowercase ): for j_focus in range(0, size_data - size_conv + 1, __lowercase ): lowercase__ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ = [] lowercase__ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__lowercase ): lowercase__ = [] for i_focus in range(len(__lowercase ) ): lowercase__ = ( np.sum(np.multiply(data_focus[i_focus], w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape( __lowercase, __lowercase ) data_featuremap.append(__lowercase ) # expanding the data slice to One dimenssion lowercase__ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__lowercase ) ) lowercase__ = np.asarray(__lowercase ) return focus_list, data_featuremap def A__ ( self : List[Any], __lowercase : Any, __lowercase : List[Any], __lowercase : Union[str, Any]="average_pool" ): # pooling process lowercase__ = len(featuremaps[0] ) lowercase__ = int(size_map / size_pooling ) lowercase__ = [] for i_map in range(len(__lowercase ) ): lowercase__ = featuremaps[i_map] lowercase__ = [] for i_focus in range(0, __lowercase, __lowercase ): for j_focus in range(0, __lowercase, __lowercase ): lowercase__ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape(__lowercase, __lowercase ) featuremap_pooled.append(__lowercase ) return featuremap_pooled def A__ ( self : str, __lowercase : Optional[Any] ): # expanding three dimension data to one dimension list lowercase__ = [] for i in range(len(__lowercase ) ): lowercase__ = np.shape(data[i] ) lowercase__ = data[i].reshape(1, shapes[0] * shapes[1] ) lowercase__ = data_listed.getA().tolist()[0] data_expanded.extend(__lowercase ) lowercase__ = np.asarray(__lowercase ) return data_expanded def A__ ( self : Optional[int], __lowercase : Optional[int] ): # expanding matrix to one dimension list lowercase__ = np.asarray(__lowercase ) lowercase__ = np.shape(__lowercase ) lowercase__ = data_mat.reshape(1, shapes[0] * shapes[1] ) return data_expanded def A__ ( self : str, __lowercase : Tuple, __lowercase : List[Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = [] lowercase__ = 0 for i_map in range(__lowercase ): lowercase__ = np.ones((size_map, size_map) ) for i in range(0, __lowercase, __lowercase ): for j in range(0, __lowercase, __lowercase ): lowercase__ = pd_pool[ i_pool ] lowercase__ = i_pool + 1 lowercase__ = np.multiply( __lowercase, np.multiply(out_map[i_map], (1 - out_map[i_map]) ) ) pd_all.append(__lowercase ) return pd_all def A__ ( self : Tuple, __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : List[str]=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(__lowercase )) ) print((" - - Shape: Teach_Data ", np.shape(__lowercase )) ) lowercase__ = 0 lowercase__ = [] lowercase__ = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase__ = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(__lowercase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ = np.asmatrix(datas_train[p] ) lowercase__ = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = np.shape(__lowercase ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = np.dot(__lowercase, self.vji.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = np.dot(__lowercase, self.wkj.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ = np.multiply( (data_teach - bp_outa), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.multiply( np.dot(__lowercase, self.wkj ), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.dot(__lowercase, self.vji ) lowercase__ = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ = pd_conva_pooled.T.getA().tolist() lowercase__ = self._calculate_gradient_from_pool( __lowercase, __lowercase, shape_featuremapa[0], shape_featuremapa[1], self.size_poolinga, ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ = self.rate_weight * np.dot(__lowercase, __lowercase ) lowercase__ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ = rp + 1 lowercase__ = error_count / patterns all_mse.append(__lowercase ) def draw_error(): lowercase__ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__lowercase, "+-" ) plt.plot(__lowercase, "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(__lowercase, alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self : List[str], __lowercase : Optional[int] ): # model predict lowercase__ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(__lowercase )) ) for p in range(len(__lowercase ) ): lowercase__ = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = bp_outa * self.vji.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = bp_outa * self.wkj.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ = [list(map(self.do_round, __lowercase ) ) for each in produce_out] return np.asarray(__lowercase ) def A__ ( self : int, __lowercase : Any ): # return the data of image after convoluting process so we can check it out lowercase__ = np.asmatrix(__lowercase ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
37
0
from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _snake_case : pass
708
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=SCREAMING_SNAKE_CASE_ , num_labels=1000 , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): lowercase__ = get_config(SCREAMING_SNAKE_CASE_ ) # load original model from timm lowercase__ = create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=SCREAMING_SNAKE_CASE_ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) lowercase__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # verify logits with torch.no_grad(): lowercase__ = model(SCREAMING_SNAKE_CASE_ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(SCREAMING_SNAKE_CASE_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE_ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowercase_ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
37
0
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class _snake_case : @property def A__ ( self : str ): return self.get_dummy_input() @property def A__ ( self : str ): if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(F'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def A__ ( self : Optional[int], __lowercase : Dict=True, __lowercase : List[Any]=False, __lowercase : Optional[int]=False, __lowercase : str=False, ): lowercase__ = 4 lowercase__ = 32 lowercase__ = (32, 32) lowercase__ = torch.manual_seed(0 ) lowercase__ = torch.device(lowerCAmelCase__ ) lowercase__ = (batch_size, num_channels) + sizes lowercase__ = randn_tensor(lowerCAmelCase__, generator=lowerCAmelCase__, device=lowerCAmelCase__ ) lowercase__ = {"hidden_states": hidden_states} if include_temb: lowercase__ = 128 lowercase__ = randn_tensor((batch_size, temb_channels), generator=lowerCAmelCase__, device=lowerCAmelCase__ ) if include_res_hidden_states_tuple: lowercase__ = torch.manual_seed(1 ) lowercase__ = (randn_tensor(lowerCAmelCase__, generator=lowerCAmelCase__, device=lowerCAmelCase__ ),) if include_encoder_hidden_states: lowercase__ = floats_tensor((batch_size, 32, 32) ).to(lowerCAmelCase__ ) if include_skip_sample: lowercase__ = randn_tensor(((batch_size, 3) + sizes), generator=lowerCAmelCase__, device=lowerCAmelCase__ ) return dummy_input def A__ ( self : str ): lowercase__ = { "in_channels": 32, "out_channels": 32, "temb_channels": 128, } if self.block_type == "up": lowercase__ = 32 if self.block_type == "mid": init_dict.pop("out_channels" ) lowercase__ = self.dummy_input return init_dict, inputs_dict def A__ ( self : Any, __lowercase : str ): lowercase__ = self.prepare_init_args_and_inputs_for_common() lowercase__ = self.block_class(**lowerCAmelCase__ ) unet_block.to(lowerCAmelCase__ ) unet_block.eval() with torch.no_grad(): lowercase__ = unet_block(**lowerCAmelCase__ ) if isinstance(lowerCAmelCase__, lowerCAmelCase__ ): lowercase__ = output[0] self.assertEqual(output.shape, self.output_shape ) lowercase__ = output[0, -1, -3:, -3:] lowercase__ = torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) assert torch_all_close(output_slice.flatten(), lowerCAmelCase__, atol=5e-3 ) @unittest.skipIf(torch_device == "mps", "Training is not supported in mps" ) def A__ ( self : Union[str, Any] ): lowercase__ = self.prepare_init_args_and_inputs_for_common() lowercase__ = self.block_class(**lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() lowercase__ = model(**lowerCAmelCase__ ) if isinstance(lowerCAmelCase__, lowerCAmelCase__ ): lowercase__ = output[0] lowercase__ = torch.device(lowerCAmelCase__ ) lowercase__ = randn_tensor(output.shape, device=lowerCAmelCase__ ) lowercase__ = torch.nn.functional.mse_loss(lowerCAmelCase__, lowerCAmelCase__ ) loss.backward()
709
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _snake_case ( lowercase__): def __init__( self : Optional[Any], __lowercase : str = "▁", __lowercase : bool = True, __lowercase : Union[str, AddedToken] = "<unk>", __lowercase : Union[str, AddedToken] = "</s>", __lowercase : Union[str, AddedToken] = "<pad>", ): lowercase__ = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } lowercase__ = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowercase__ = token_dict["token"] lowercase__ = Tokenizer(Unigram() ) lowercase__ = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ), " " ), normalizers.Lowercase(), ] ) lowercase__ = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) lowercase__ = decoders.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ) lowercase__ = TemplateProcessing( single=F'''$A {self.special_tokens["eos"]["token"]}''', special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])], ) lowercase__ = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase, __lowercase ) def A__ ( self : Union[str, Any], __lowercase : Union[str, List[str]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) if isinstance(__lowercase, __lowercase ): lowercase__ = [files] self._tokenizer.train(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : List[Any], __lowercase : Union[Iterator[str], Iterator[Iterator[str]]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) self._tokenizer.train_from_iterator(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : str ): lowercase__ = json.loads(self._tokenizer.to_str() ) lowercase__ = self.special_tokens["unk"]["id"] lowercase__ = Tokenizer.from_str(json.dumps(__lowercase ) )
37
0
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig lowercase_ = logging.get_logger(__name__) # General docstring lowercase_ = '''PoolFormerConfig''' # Base docstring lowercase_ = '''sail/poolformer_s12''' lowercase_ = [1, 512, 7, 7] # Image classification docstring lowercase_ = '''sail/poolformer_s12''' lowercase_ = '''tabby, tabby cat''' lowercase_ = [ '''sail/poolformer_s12''', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = False ): if drop_prob == 0.0 or not training: return input lowercase__ = 1 - drop_prob lowercase__ = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets lowercase__ = keep_prob + torch.rand(UpperCAmelCase__ , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize lowercase__ = input.div(UpperCAmelCase__ ) * random_tensor return output class _snake_case ( nn.Module): def __init__( self : Tuple, __lowercase : Optional[float] = None ): super().__init__() lowercase__ = drop_prob def A__ ( self : Union[str, Any], __lowercase : torch.Tensor ): return drop_path(__lowerCAmelCase, self.drop_prob, self.training ) def A__ ( self : List[str] ): return "p={}".format(self.drop_prob ) class _snake_case ( nn.Module): def __init__( self : Union[str, Any], __lowercase : Tuple, __lowercase : Union[str, Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Dict, __lowercase : Dict=None ): super().__init__() lowercase__ = patch_size if isinstance(__lowerCAmelCase, collections.abc.Iterable ) else (patch_size, patch_size) lowercase__ = stride if isinstance(__lowerCAmelCase, collections.abc.Iterable ) else (stride, stride) lowercase__ = padding if isinstance(__lowerCAmelCase, collections.abc.Iterable ) else (padding, padding) lowercase__ = nn.Convad(__lowerCAmelCase, __lowerCAmelCase, kernel_size=__lowerCAmelCase, stride=__lowerCAmelCase, padding=__lowerCAmelCase ) lowercase__ = norm_layer(__lowerCAmelCase ) if norm_layer else nn.Identity() def A__ ( self : List[str], __lowercase : Union[str, Any] ): lowercase__ = self.projection(__lowerCAmelCase ) lowercase__ = self.norm(__lowerCAmelCase ) return embeddings class _snake_case ( nn.GroupNorm): def __init__( self : Tuple, __lowercase : int, **__lowercase : Optional[Any] ): super().__init__(1, __lowerCAmelCase, **__lowerCAmelCase ) class _snake_case ( nn.Module): def __init__( self : int, __lowercase : Optional[Any] ): super().__init__() lowercase__ = nn.AvgPoolad(__lowerCAmelCase, stride=1, padding=pool_size // 2, count_include_pad=__lowerCAmelCase ) def A__ ( self : Optional[Any], __lowercase : List[Any] ): return self.pool(__lowerCAmelCase ) - hidden_states class _snake_case ( nn.Module): def __init__( self : Tuple, __lowercase : int, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Any ): super().__init__() lowercase__ = nn.Convad(__lowerCAmelCase, __lowerCAmelCase, 1 ) lowercase__ = nn.Convad(__lowerCAmelCase, __lowerCAmelCase, 1 ) lowercase__ = PoolFormerDropPath(__lowerCAmelCase ) if isinstance(config.hidden_act, __lowerCAmelCase ): lowercase__ = ACTaFN[config.hidden_act] else: lowercase__ = config.hidden_act def A__ ( self : Tuple, __lowercase : List[str] ): lowercase__ = self.conva(__lowerCAmelCase ) lowercase__ = self.act_fn(__lowerCAmelCase ) lowercase__ = self.drop(__lowerCAmelCase ) lowercase__ = self.conva(__lowerCAmelCase ) lowercase__ = self.drop(__lowerCAmelCase ) return hidden_states class _snake_case ( nn.Module): def __init__( self : Dict, __lowercase : Optional[Any], __lowercase : int, __lowercase : Dict, __lowercase : Dict, __lowercase : Tuple, __lowercase : Union[str, Any] ): super().__init__() lowercase__ = PoolFormerPooling(__lowerCAmelCase ) lowercase__ = PoolFormerOutput(__lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase ) lowercase__ = PoolFormerGroupNorm(__lowerCAmelCase ) lowercase__ = PoolFormerGroupNorm(__lowerCAmelCase ) # Useful for training neural nets lowercase__ = PoolFormerDropPath(__lowerCAmelCase ) if drop_path > 0.0 else nn.Identity() lowercase__ = config.use_layer_scale if config.use_layer_scale: lowercase__ = nn.Parameter( config.layer_scale_init_value * torch.ones((__lowerCAmelCase) ), requires_grad=__lowerCAmelCase ) lowercase__ = nn.Parameter( config.layer_scale_init_value * torch.ones((__lowerCAmelCase) ), requires_grad=__lowerCAmelCase ) def A__ ( self : Dict, __lowercase : Dict ): if self.use_layer_scale: lowercase__ = self.pooling(self.before_norm(__lowerCAmelCase ) ) lowercase__ = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection lowercase__ = hidden_states + self.drop_path(__lowerCAmelCase ) lowercase__ = () lowercase__ = self.output(self.after_norm(__lowerCAmelCase ) ) lowercase__ = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection lowercase__ = hidden_states + self.drop_path(__lowerCAmelCase ) lowercase__ = (output,) + outputs return outputs else: lowercase__ = self.drop_path(self.pooling(self.before_norm(__lowerCAmelCase ) ) ) # First residual connection lowercase__ = pooling_output + hidden_states lowercase__ = () # Second residual connection inside the PoolFormerOutput block lowercase__ = self.drop_path(self.output(self.after_norm(__lowerCAmelCase ) ) ) lowercase__ = hidden_states + layer_output lowercase__ = (output,) + outputs return outputs class _snake_case ( nn.Module): def __init__( self : Any, __lowercase : Dict ): super().__init__() lowercase__ = config # stochastic depth decay rule lowercase__ = [x.item() for x in torch.linspace(0, config.drop_path_rate, sum(config.depths ) )] # patch embeddings lowercase__ = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i], stride=config.strides[i], padding=config.padding[i], num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1], hidden_size=config.hidden_sizes[i], ) ) lowercase__ = nn.ModuleList(__lowerCAmelCase ) # Transformer blocks lowercase__ = [] lowercase__ = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers lowercase__ = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( __lowerCAmelCase, num_channels=config.hidden_sizes[i], pool_size=config.pool_size, hidden_size=config.hidden_sizes[i], intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ), drop_path=dpr[cur + j], ) ) blocks.append(nn.ModuleList(__lowerCAmelCase ) ) lowercase__ = nn.ModuleList(__lowerCAmelCase ) def A__ ( self : Any, __lowercase : Dict, __lowercase : Optional[Any]=False, __lowercase : List[str]=True ): lowercase__ = () if output_hidden_states else None lowercase__ = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings, self.block ) ): lowercase__ , lowercase__ = layers # Get patch embeddings from hidden_states lowercase__ = embedding_layer(__lowerCAmelCase ) # Send the embeddings through the blocks for _, blk in enumerate(__lowerCAmelCase ): lowercase__ = blk(__lowerCAmelCase ) lowercase__ = layer_outputs[0] if output_hidden_states: lowercase__ = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=__lowerCAmelCase, hidden_states=__lowerCAmelCase ) class _snake_case ( UpperCAmelCase__): UpperCamelCase__ : Any =PoolFormerConfig UpperCamelCase__ : List[Any] ="""poolformer""" UpperCamelCase__ : int ="""pixel_values""" UpperCamelCase__ : Optional[int] =True def A__ ( self : List[str], __lowercase : List[str] ): if isinstance(__lowerCAmelCase, (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0, std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(__lowerCAmelCase, nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def A__ ( self : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any]=False ): if isinstance(__lowerCAmelCase, __lowerCAmelCase ): lowercase__ = value lowercase_ = r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' lowercase_ = r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`PoolFormerImageProcessor.__call__`] for details. ''' @add_start_docstrings( """The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top.""" , UpperCAmelCase__ , ) class _snake_case ( UpperCAmelCase__): def __init__( self : Any, __lowercase : Union[str, Any] ): super().__init__(__lowerCAmelCase ) lowercase__ = config lowercase__ = PoolFormerEncoder(__lowerCAmelCase ) # Initialize weights and apply final processing self.post_init() def A__ ( self : List[Any] ): return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(__lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC, output_type=__lowerCAmelCase, config_class=_CONFIG_FOR_DOC, modality="vision", expected_output=_EXPECTED_OUTPUT_SHAPE, ) def A__ ( self : int, __lowercase : Optional[torch.FloatTensor] = None, __lowercase : Optional[bool] = None, __lowercase : Optional[bool] = None, ): lowercase__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__ = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) lowercase__ = self.encoder( __lowerCAmelCase, output_hidden_states=__lowerCAmelCase, return_dict=__lowerCAmelCase, ) lowercase__ = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=__lowerCAmelCase, hidden_states=encoder_outputs.hidden_states, ) class _snake_case ( nn.Module): def __init__( self : Union[str, Any], __lowercase : Any ): super().__init__() lowercase__ = nn.Linear(config.hidden_size, config.hidden_size ) def A__ ( self : int, __lowercase : List[Any] ): lowercase__ = self.dense(__lowerCAmelCase ) return output @add_start_docstrings( """ PoolFormer Model transformer with an image classification head on top """ , UpperCAmelCase__ , ) class _snake_case ( UpperCAmelCase__): def __init__( self : Dict, __lowercase : Optional[int] ): super().__init__(__lowerCAmelCase ) lowercase__ = config.num_labels lowercase__ = PoolFormerModel(__lowerCAmelCase ) # Final norm lowercase__ = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head lowercase__ = ( nn.Linear(config.hidden_sizes[-1], config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__lowerCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT, output_type=__lowerCAmelCase, config_class=_CONFIG_FOR_DOC, expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT, ) def A__ ( self : Union[str, Any], __lowercase : Optional[torch.FloatTensor] = None, __lowercase : Optional[torch.LongTensor] = None, __lowercase : Optional[bool] = None, __lowercase : Optional[bool] = None, ): lowercase__ = return_dict if return_dict is not None else self.config.use_return_dict lowercase__ = self.poolformer( __lowerCAmelCase, output_hidden_states=__lowerCAmelCase, return_dict=__lowerCAmelCase, ) lowercase__ = outputs[0] lowercase__ = self.classifier(self.norm(__lowerCAmelCase ).mean([-2, -1] ) ) lowercase__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase__ = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase__ = "single_label_classification" else: lowercase__ = "multi_label_classification" if self.config.problem_type == "regression": lowercase__ = MSELoss() if self.num_labels == 1: lowercase__ = loss_fct(logits.squeeze(), labels.squeeze() ) else: lowercase__ = loss_fct(__lowerCAmelCase, __lowerCAmelCase ) elif self.config.problem_type == "single_label_classification": lowercase__ = CrossEntropyLoss() lowercase__ = loss_fct(logits.view(-1, self.num_labels ), labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase__ = BCEWithLogitsLoss() lowercase__ = loss_fct(__lowerCAmelCase, __lowerCAmelCase ) if not return_dict: lowercase__ = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=__lowerCAmelCase, logits=__lowerCAmelCase, hidden_states=outputs.hidden_states )
710
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase__ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowercase__ = f'''{src_lang}-{tgt_lang}''' lowercase__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) lowercase__ = os.path.join(SCREAMING_SNAKE_CASE_ , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
37
0
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class _snake_case ( __UpperCAmelCase): UpperCamelCase__ : Union[str, Any] =(CMStochasticIterativeScheduler,) UpperCamelCase__ : str =1_0 def A__ ( self : Optional[int], **__lowercase : Optional[Any] ): lowercase__ = { "num_train_timesteps": 201, "sigma_min": 0.002, "sigma_max": 80.0, } config.update(**UpperCAmelCase_ ) return config def A__ ( self : Optional[int] ): lowercase__ = 10 lowercase__ = self.get_scheduler_config() lowercase__ = self.scheduler_classes[0](**UpperCAmelCase_ ) scheduler.set_timesteps(UpperCAmelCase_ ) lowercase__ = scheduler.timesteps[0] lowercase__ = scheduler.timesteps[1] lowercase__ = self.dummy_sample lowercase__ = 0.1 * sample lowercase__ = scheduler.step(UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_ ).prev_sample lowercase__ = scheduler.step(UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_ ).prev_sample self.assertEqual(output_a.shape, sample.shape ) self.assertEqual(output_a.shape, output_a.shape ) def A__ ( self : str ): for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase_ ) def A__ ( self : int ): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=UpperCAmelCase_ ) def A__ ( self : Any ): lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**UpperCAmelCase_ ) lowercase__ = 1 scheduler.set_timesteps(UpperCAmelCase_ ) lowercase__ = scheduler.timesteps lowercase__ = torch.manual_seed(0 ) lowercase__ = self.dummy_model() lowercase__ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(UpperCAmelCase_ ): # 1. scale model input lowercase__ = scheduler.scale_model_input(UpperCAmelCase_, UpperCAmelCase_ ) # 2. predict noise residual lowercase__ = model(UpperCAmelCase_, UpperCAmelCase_ ) # 3. predict previous sample x_t-1 lowercase__ = scheduler.step(UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_, generator=UpperCAmelCase_ ).prev_sample lowercase__ = pred_prev_sample lowercase__ = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowercase__ = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 192.7614 ) < 1e-2 assert abs(result_mean.item() - 0.2510 ) < 1e-3 def A__ ( self : Dict ): lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**UpperCAmelCase_ ) lowercase__ = [106, 0] scheduler.set_timesteps(timesteps=UpperCAmelCase_ ) lowercase__ = scheduler.timesteps lowercase__ = torch.manual_seed(0 ) lowercase__ = self.dummy_model() lowercase__ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input lowercase__ = scheduler.scale_model_input(UpperCAmelCase_, UpperCAmelCase_ ) # 2. predict noise residual lowercase__ = model(UpperCAmelCase_, UpperCAmelCase_ ) # 3. predict previous sample x_t-1 lowercase__ = scheduler.step(UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_, generator=UpperCAmelCase_ ).prev_sample lowercase__ = pred_prev_sample lowercase__ = torch.sum(torch.abs(UpperCAmelCase_ ) ) lowercase__ = torch.mean(torch.abs(UpperCAmelCase_ ) ) assert abs(result_sum.item() - 347.6357 ) < 1e-2 assert abs(result_mean.item() - 0.4527 ) < 1e-3 def A__ ( self : Dict ): lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**UpperCAmelCase_ ) lowercase__ = [39, 30, 12, 15, 0] with self.assertRaises(UpperCAmelCase_, msg="`timesteps` must be in descending order." ): scheduler.set_timesteps(timesteps=UpperCAmelCase_ ) def A__ ( self : Optional[int] ): lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**UpperCAmelCase_ ) lowercase__ = [39, 30, 12, 1, 0] lowercase__ = len(UpperCAmelCase_ ) with self.assertRaises(UpperCAmelCase_, msg="Can only pass one of `num_inference_steps` or `timesteps`." ): scheduler.set_timesteps(num_inference_steps=UpperCAmelCase_, timesteps=UpperCAmelCase_ ) def A__ ( self : Any ): lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**UpperCAmelCase_ ) lowercase__ = [scheduler.config.num_train_timesteps] with self.assertRaises( UpperCAmelCase_, msg="`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}", ): scheduler.set_timesteps(timesteps=UpperCAmelCase_ )
711
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =TransfoXLTokenizer UpperCamelCase__ : List[Any] =False UpperCamelCase__ : List[Any] =False def A__ ( self : Union[str, Any] ): super().setUp() lowercase__ = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : Union[str, Any], **__lowercase : Any ): lowercase__ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname, **__lowercase ) def A__ ( self : Tuple, __lowercase : Optional[int] ): lowercase__ = "<unk> UNwanted , running" lowercase__ = "<unk> unwanted, running" return input_text, output_text def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(vocab_file=self.vocab_file, lower_case=__lowercase ) lowercase__ = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowercase, ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ), [0, 4, 8, 7] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["hello", "!", "how", "are", "you", "?"] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) lowercase__ = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" lowercase__ = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowercase ), __lowercase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowercase ), __lowercase ) def A__ ( self : List[str] ): lowercase__ = self.get_tokenizer() lowercase__ = len(__lowercase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1", 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowercase ), original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ), [1] ) self.assertEqual(tokenizer.decode([1] ), "new1" )
37
0
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand lowercase_ = logging.get_logger(__name__) # pylint: disable=invalid-name def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(_lowercase ): return ext raise Exception( f'''Unable to determine file format from file extension {path}. ''' f'''Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}''' ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) lowercase__ = try_infer_format_from_ext(args.input ) if args.format == '''infer''' else args.format lowercase__ = PipelineDataFormat.from_str( format=_lowercase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(_lowercase , _lowercase ) class _snake_case ( UpperCamelCase_): def __init__( self : Optional[Any], __lowercase : Pipeline, __lowercase : PipelineDataFormat ): lowercase__ = nlp lowercase__ = reader @staticmethod def A__ ( __lowercase : ArgumentParser ): lowercase__ = parser.add_parser("run", help="Run a pipeline through the CLI" ) run_parser.add_argument("--task", choices=get_supported_tasks(), help="Task to run" ) run_parser.add_argument("--input", type=UpperCamelCase__, help="Path to the file to use for inference" ) run_parser.add_argument("--output", type=UpperCamelCase__, help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model", type=UpperCamelCase__, help="Name or path to the model to instantiate." ) run_parser.add_argument("--config", type=UpperCamelCase__, help="Name or path to the model\'s config to instantiate." ) run_parser.add_argument( "--tokenizer", type=UpperCamelCase__, help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column", type=UpperCamelCase__, help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)", ) run_parser.add_argument( "--format", type=UpperCamelCase__, default="infer", choices=PipelineDataFormat.SUPPORTED_FORMATS, help="Input format to read from", ) run_parser.add_argument( "--device", type=UpperCamelCase__, default=-1, help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)", ) run_parser.add_argument("--overwrite", action="store_true", help="Allow overwriting the output file." ) run_parser.set_defaults(func=UpperCamelCase__ ) def A__ ( self : Optional[Any] ): lowercase__ = self._nlp, [] for entry in self._reader: lowercase__ = nlp(**UpperCamelCase__ ) if self._reader.is_multi_columns else nlp(UpperCamelCase__ ) if isinstance(UpperCamelCase__, UpperCamelCase__ ): outputs.append(UpperCamelCase__ ) else: outputs += output # Saving data if self._nlp.binary_output: lowercase__ = self._reader.save_binary(UpperCamelCase__ ) logger.warning(F'''Current pipeline requires output to be in binary format, saving at {binary_path}''' ) else: self._reader.save(UpperCamelCase__ )
712
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser(SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.parse_args_into_dataclasses()[0] lowercase__ = TensorFlowBenchmark(args=SCREAMING_SNAKE_CASE_ ) try: lowercase__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowercase__ = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowercase__ = " ".join(str(SCREAMING_SNAKE_CASE_ ).split(" " )[:-1] ) lowercase__ = "" lowercase__ = eval(str(SCREAMING_SNAKE_CASE_ ).split(" " )[-1] ) lowercase__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = full_error_msg + begin_error_msg + str(SCREAMING_SNAKE_CASE_ ) raise ValueError(SCREAMING_SNAKE_CASE_ ) benchmark.run() if __name__ == "__main__": main()
37
0
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class _snake_case ( lowercase__): def A__ ( self : List[Any] ): return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def A__ ( self : Dict ): lowercase__ = {"col_1": [3, 2, 1, 0], "col_2": ["a", "b", "c", "d"]} return Dataset.from_dict(_SCREAMING_SNAKE_CASE ) def A__ ( self : str ): lowercase__ = self._create_example_records() lowercase__ = Dataset.from_list(_SCREAMING_SNAKE_CASE ) self.assertListEqual(dset.column_names, ["col_1", "col_2"] ) for i, r in enumerate(_SCREAMING_SNAKE_CASE ): self.assertDictEqual(_SCREAMING_SNAKE_CASE, example_records[i] ) def A__ ( self : Tuple ): lowercase__ = self._create_example_records() lowercase__ = Dataset.from_list(_SCREAMING_SNAKE_CASE ) lowercase__ = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info, dset_from_dict.info ) def A__ ( self : Tuple ): # checks what happens with missing columns lowercase__ = [{"col_1": 1}, {"col_2": "x"}] lowercase__ = Dataset.from_list(_SCREAMING_SNAKE_CASE ) self.assertDictEqual(dset[0], {"col_1": 1} ) self.assertDictEqual(dset[1], {"col_1": None} ) # NB: first record is used for columns def A__ ( self : Union[str, Any] ): # checks if the type can be inferred from the second record lowercase__ = [{"col_1": []}, {"col_1": [1, 2]}] lowercase__ = Dataset.from_list(_SCREAMING_SNAKE_CASE ) self.assertEqual(dset.info.features["col_1"], Sequence(Value("int64" ) ) ) def A__ ( self : Tuple ): lowercase__ = Dataset.from_list([] ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ), 0 ) self.assertListEqual(dset.column_names, [] )
713
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowercase_ = """<<<<<<< This should probably be modified because it mentions: """ lowercase_ = """======= >>>>>>> """ lowercase_ = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] lowercase_ = [ # (pattern, replacement) # Order is important here for some replacements (r"""tfds\.core""", r"""datasets"""), (r"""tf\.io\.gfile\.GFile""", r"""open"""), (r"""tf\.([\w\d]+)""", r"""datasets.Value('\1')"""), (r"""tfds\.features\.Text\(\)""", r"""datasets.Value('string')"""), (r"""tfds\.features\.Text\(""", r"""datasets.Value('string'),"""), (r"""features\s*=\s*tfds.features.FeaturesDict\(""", r"""features=datasets.Features("""), (r"""tfds\.features\.FeaturesDict\(""", r"""dict("""), (r"""The TensorFlow Datasets Authors""", r"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (r"""tfds\.""", r"""datasets."""), (r"""dl_manager\.manual_dir""", r"""self.config.data_dir"""), (r"""self\.builder_config""", r"""self.config"""), ] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return ConvertCommand(args.tfds_path , args.datasets_directory ) class _snake_case ( lowercase__): @staticmethod def A__ ( __lowercase : ArgumentParser ): lowercase__ = parser.add_parser( "convert", help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.", ) train_parser.add_argument( "--tfds_path", type=__lowercase, required=__lowercase, help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.", ) train_parser.add_argument( "--datasets_directory", type=__lowercase, required=__lowercase, help="Path to the HuggingFace Datasets folder." ) train_parser.set_defaults(func=__lowercase ) def __init__( self : Tuple, __lowercase : str, __lowercase : str, *__lowercase : Tuple ): lowercase__ = get_logger("datasets-cli/converting" ) lowercase__ = tfds_path lowercase__ = datasets_directory def A__ ( self : Any ): if os.path.isdir(self._tfds_path ): lowercase__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__ = os.path.dirname(self._tfds_path ) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path." ) lowercase__ = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) lowercase__ = [] lowercase__ = [] lowercase__ = {} if os.path.isdir(self._tfds_path ): lowercase__ = os.listdir(__lowercase ) else: lowercase__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) if not os.path.isfile(__lowercase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file" ) continue with open(__lowercase, encoding="utf-8" ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = False lowercase__ = False lowercase__ = [] for line in lines: lowercase__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__ = "import datasets\n" elif "import tensorflow" in out_line: # order is important here lowercase__ = "" continue elif "from absl import logging" in out_line: lowercase__ = "from datasets import logging\n" elif "getLogger" in out_line: lowercase__ = out_line.replace("getLogger", "get_logger" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__ = True lowercase__ = list(filter(lambda __lowercase : e in out_line, __lowercase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowercase ) + "\n" ) out_lines.append(__lowercase ) out_lines.append(__lowercase ) continue else: for pattern, replacement in TO_CONVERT: lowercase__ = re.sub(__lowercase, __lowercase, __lowercase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__ = re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)", __lowercase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split("," ) ) lowercase__ = "from . import " + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__ = True out_lines.append(__lowercase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__ = f_name.replace(".py", "" ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) os.makedirs(__lowercase, exist_ok=__lowercase ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowercase ) if needs_manual_update: with_manual_update.append(__lowercase ) with open(__lowercase, "w", encoding="utf-8" ) as f: f.writelines(__lowercase ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: lowercase__ = os.path.basename(__lowercase ) lowercase__ = imports_to_builder_map[f_name.replace(".py", "" )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowercase, __lowercase ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
37
0
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _snake_case ( pl.LightningModule): def __init__( self : List[Any], __lowercase : List[str] ): super().__init__() lowercase__ = model lowercase__ = 2 lowercase__ = nn.Linear(self.model.config.hidden_size, self.num_labels ) def A__ ( self : Optional[Any] ): pass def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): # load longformer model from model identifier lowercase__ = LongformerModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowercase__ = LightningModel(SCREAMING_SNAKE_CASE_ ) lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model lowercase__ = LongformerForQuestionAnswering.from_pretrained(SCREAMING_SNAKE_CASE_ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(f'''Conversion successful. Model saved under {pytorch_dump_folder_path}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase_ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
714
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase_ = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } lowercase_ = { """allenai/led-base-16384""": 1_6384, } class _snake_case ( lowercase__): UpperCamelCase__ : int =VOCAB_FILES_NAMES UpperCamelCase__ : Any =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =LEDTokenizer UpperCamelCase__ : Tuple =["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any], __lowercase : Optional[Any]=None, __lowercase : Dict=None, __lowercase : Tuple=None, __lowercase : Union[str, Any]="replace", __lowercase : Tuple="<s>", __lowercase : Optional[Any]="</s>", __lowercase : Tuple="</s>", __lowercase : List[str]="<s>", __lowercase : Tuple="<unk>", __lowercase : Dict="<pad>", __lowercase : Dict="<mask>", __lowercase : Any=False, __lowercase : Any=True, **__lowercase : List[Any], ): super().__init__( __lowercase, __lowercase, tokenizer_file=__lowercase, errors=__lowercase, bos_token=__lowercase, eos_token=__lowercase, sep_token=__lowercase, cls_token=__lowercase, unk_token=__lowercase, pad_token=__lowercase, mask_token=__lowercase, add_prefix_space=__lowercase, trim_offsets=__lowercase, **__lowercase, ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = getattr(__lowercase, pre_tok_state.pop("type" ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**__lowercase ) lowercase__ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase__ = "post_processor" lowercase__ = getattr(self.backend_tokenizer, __lowercase, __lowercase ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state["sep"] ) if "cls" in state: lowercase__ = tuple(state["cls"] ) lowercase__ = False if state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get("trim_offsets", __lowercase ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(__lowercase, state.pop("type" ) ) lowercase__ = component_class(**__lowercase ) setattr(self.backend_tokenizer, __lowercase, __lowercase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A__ ( self : str ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self : Optional[int], __lowercase : Dict ): lowercase__ = AddedToken(__lowercase, lstrip=__lowercase, rstrip=__lowercase ) if isinstance(__lowercase, __lowercase ) else value lowercase__ = value def A__ ( self : Any, *__lowercase : List[Any], **__lowercase : Optional[Any] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowercase, **__lowercase ) def A__ ( self : int, *__lowercase : Union[str, Any], **__lowercase : List[str] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowercase, **__lowercase ) def A__ ( self : Optional[Any], __lowercase : str, __lowercase : Optional[str] = None ): lowercase__ = self._tokenizer.model.save(__lowercase, name=__lowercase ) return tuple(__lowercase ) def A__ ( self : List[str], __lowercase : int, __lowercase : Optional[int]=None ): lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self : int, __lowercase : List[int], __lowercase : Optional[List[int]] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self : Union[str, Any], __lowercase : Union[Dict[str, EncodedInput], BatchEncoding], __lowercase : Optional[int] = None, __lowercase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD, __lowercase : Optional[int] = None, __lowercase : Optional[bool] = None, ): lowercase__ = super()._pad( encoded_inputs=__lowercase, max_length=__lowercase, padding_strategy=__lowercase, pad_to_multiple_of=__lowercase, return_attention_mask=__lowercase, ) # Load from model defaults if return_attention_mask is None: lowercase__ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ = len(encoded_inputs["global_attention_mask"] ) != len(__lowercase ) if needs_to_be_padded: lowercase__ = len(__lowercase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
37
0
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } lowercase_ = { """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } lowercase_ = { """vinai/phobert-base""": 256, """vinai/phobert-large""": 256, } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = set() lowercase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ = char lowercase__ = set(_lowerCAmelCase ) return pairs class _snake_case ( __snake_case): UpperCamelCase__ : Optional[int] =VOCAB_FILES_NAMES UpperCamelCase__ : str =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int], __lowercase : List[str], __lowercase : Optional[int], __lowercase : Union[str, Any]="<s>", __lowercase : List[Any]="</s>", __lowercase : List[str]="</s>", __lowercase : Optional[Any]="<s>", __lowercase : int="<unk>", __lowercase : Optional[Any]="<pad>", __lowercase : Optional[Any]="<mask>", **__lowercase : str, ): super().__init__( bos_token=A_, eos_token=A_, unk_token=A_, sep_token=A_, cls_token=A_, pad_token=A_, mask_token=A_, **A_, ) lowercase__ = vocab_file lowercase__ = merges_file lowercase__ = {} lowercase__ = 0 lowercase__ = 1 lowercase__ = 2 lowercase__ = 3 self.add_from_file(A_ ) lowercase__ = {v: k for k, v in self.encoder.items()} with open(A_, encoding="utf-8" ) as merges_handle: lowercase__ = merges_handle.read().split("\n" )[:-1] lowercase__ = [tuple(merge.split()[:-1] ) for merge in merges] lowercase__ = dict(zip(A_, range(len(A_ ) ) ) ) lowercase__ = {} def A__ ( self : Union[str, Any], __lowercase : str, __lowercase : List[str] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self : List[Any], __lowercase : List[str], __lowercase : int = None, __lowercase : Dict = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_, token_ids_a=A_, already_has_special_tokens=A_ ) if token_ids_a is None: return [1] + ([0] * len(A_ )) + [1] return [1] + ([0] * len(A_ )) + [1, 1] + ([0] * len(A_ )) + [1] def A__ ( self : Dict, __lowercase : int, __lowercase : Optional[int] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A__ ( self : Optional[Any] ): return len(self.encoder ) def A__ ( self : Dict ): return dict(self.encoder, **self.added_tokens_encoder ) def A__ ( self : Union[str, Any], __lowercase : List[str] ): if token in self.cache: return self.cache[token] lowercase__ = tuple(A_ ) lowercase__ = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) lowercase__ = get_pairs(A_ ) if not pairs: return token while True: lowercase__ = min(A_, key=lambda __lowercase : self.bpe_ranks.get(A_, float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowercase__ = bigram lowercase__ = [] lowercase__ = 0 while i < len(A_ ): try: lowercase__ = word.index(A_, A_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ = j if word[i] == first and i < len(A_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ = tuple(A_ ) lowercase__ = new_word if len(A_ ) == 1: break else: lowercase__ = get_pairs(A_ ) lowercase__ = "@@ ".join(A_ ) lowercase__ = word[:-4] lowercase__ = word return word def A__ ( self : Tuple, __lowercase : Any ): lowercase__ = [] lowercase__ = re.findall(R"\S+\n?", A_ ) for token in words: split_tokens.extend(list(self.bpe(A_ ).split(" " ) ) ) return split_tokens def A__ ( self : List[Any], __lowercase : Any ): return self.encoder.get(A_, self.encoder.get(self.unk_token ) ) def A__ ( self : Optional[Any], __lowercase : int ): return self.decoder.get(A_, self.unk_token ) def A__ ( self : List[Any], __lowercase : Tuple ): lowercase__ = " ".join(A_ ).replace("@@ ", "" ).strip() return out_string def A__ ( self : List[Any], __lowercase : Dict, __lowercase : int = None ): if not os.path.isdir(A_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( A_, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ = os.path.join( A_, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ): copyfile(self.vocab_file, A_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(A_ ): copyfile(self.merges_file, A_ ) return out_vocab_file, out_merge_file def A__ ( self : Optional[int], __lowercase : List[str] ): if isinstance(A_, A_ ): try: with open(A_, "r", encoding="utf-8" ) as fd: self.add_from_file(A_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return lowercase__ = f.readlines() for lineTmp in lines: lowercase__ = lineTmp.strip() lowercase__ = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) lowercase__ = line[:idx] lowercase__ = len(self.encoder )
715
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __lowerCAmelCase ( ): lowercase__ = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=SCREAMING_SNAKE_CASE_ ) env_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) launch_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) tpu_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) test_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) # Let's go lowercase__ = parser.parse_args() if not hasattr(SCREAMING_SNAKE_CASE_ , "func" ): parser.print_help() exit(1 ) # Run args.func(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
37
0
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class _snake_case ( UpperCamelCase__): UpperCamelCase__ : torch.FloatTensor class _snake_case ( UpperCamelCase__ , UpperCamelCase__): @register_to_config def __init__( self : List[str], __lowercase : Optional[int] = 16, __lowercase : Any = 88, __lowercase : Union[str, Any] = None, __lowercase : str = None, __lowercase : Optional[Any] = 1, __lowercase : Optional[int] = 0.0, __lowercase : Optional[Any] = 32, __lowercase : Optional[Any] = None, __lowercase : Dict = False, __lowercase : Any = None, __lowercase : List[Any] = "geglu", __lowercase : Optional[Any] = True, __lowercase : Optional[Any] = True, ): super().__init__() lowercase__ = num_attention_heads lowercase__ = attention_head_dim lowercase__ = num_attention_heads * attention_head_dim lowercase__ = in_channels lowercase__ = torch.nn.GroupNorm(num_groups=__lowercase, num_channels=__lowercase, eps=1e-6, affine=__lowercase ) lowercase__ = nn.Linear(__lowercase, __lowercase ) # 3. Define transformers blocks lowercase__ = nn.ModuleList( [ BasicTransformerBlock( __lowercase, __lowercase, __lowercase, dropout=__lowercase, cross_attention_dim=__lowercase, activation_fn=__lowercase, attention_bias=__lowercase, double_self_attention=__lowercase, norm_elementwise_affine=__lowercase, ) for d in range(__lowercase ) ] ) lowercase__ = nn.Linear(__lowercase, __lowercase ) def A__ ( self : List[Any], __lowercase : Union[str, Any], __lowercase : Optional[int]=None, __lowercase : Optional[int]=None, __lowercase : str=None, __lowercase : Tuple=1, __lowercase : Optional[Any]=None, __lowercase : List[Any] = True, ): lowercase__ , lowercase__ , lowercase__ , lowercase__ = hidden_states.shape lowercase__ = batch_frames // num_frames lowercase__ = hidden_states lowercase__ = hidden_states[None, :].reshape(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) lowercase__ = hidden_states.permute(0, 2, 1, 3, 4 ) lowercase__ = self.norm(__lowercase ) lowercase__ = hidden_states.permute(0, 3, 4, 2, 1 ).reshape(batch_size * height * width, __lowercase, __lowercase ) lowercase__ = self.proj_in(__lowercase ) # 2. Blocks for block in self.transformer_blocks: lowercase__ = block( __lowercase, encoder_hidden_states=__lowercase, timestep=__lowercase, cross_attention_kwargs=__lowercase, class_labels=__lowercase, ) # 3. Output lowercase__ = self.proj_out(__lowercase ) lowercase__ = ( hidden_states[None, None, :] .reshape(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) .permute(0, 3, 4, 1, 2 ) .contiguous() ) lowercase__ = hidden_states.reshape(__lowercase, __lowercase, __lowercase, __lowercase ) lowercase__ = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=__lowercase )
716
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : Dict, __lowercase : int, __lowercase : Union[str, Any]=7, __lowercase : Union[str, Any]=3, __lowercase : Any=18, __lowercase : Union[str, Any]=30, __lowercase : Any=400, __lowercase : List[str]=True, __lowercase : Dict=None, __lowercase : List[str]=True, __lowercase : int=False, __lowercase : Union[str, Any]=True, __lowercase : str=True, __lowercase : Optional[int]=[0.5, 0.5, 0.5], __lowercase : List[Any]=[0.5, 0.5, 0.5], ): lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size if size is not None else {"height": 18, "width": 20} lowercase__ = do_thumbnail lowercase__ = do_align_axis lowercase__ = do_pad lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std def A__ ( self : Optional[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =DonutImageProcessor if is_vision_available() else None def A__ ( self : str ): lowercase__ = DonutImageProcessingTester(self ) @property def A__ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_resize" ) ) self.assertTrue(hasattr(__lowercase, "size" ) ) self.assertTrue(hasattr(__lowercase, "do_thumbnail" ) ) self.assertTrue(hasattr(__lowercase, "do_align_long_axis" ) ) self.assertTrue(hasattr(__lowercase, "do_pad" ) ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "image_mean" ) ) self.assertTrue(hasattr(__lowercase, "image_std" ) ) def A__ ( self : str ): lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {"height": 18, "width": 20} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {"height": 42, "width": 42} ) # Previous config had dimensions in (width, height) order lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=(42, 84) ) self.assertEqual(image_processor.size, {"height": 84, "width": 42} ) def A__ ( self : List[str] ): pass @is_flaky() def A__ ( self : Dict ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Optional[Any] ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Tuple ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), )
37
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase_ = {"""configuration_plbart""": ["""PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PLBartConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["""PLBartTokenizer"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """PLBART_PRETRAINED_MODEL_ARCHIVE_LIST""", """PLBartForCausalLM""", """PLBartForConditionalGeneration""", """PLBartForSequenceClassification""", """PLBartModel""", """PLBartPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
717
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _snake_case ( lowercase__): def A__ ( self : Optional[Any], __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) lowercase__ = input_file.read() lowercase__ = regexp.search(__lowercase ) return match def A__ ( self : str, __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()", re.DOTALL ) lowercase__ = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowercase__ = regexp.finditer(__lowercase ) lowercase__ = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(F'''open(...) must use utf-8 encoding in {dataset}''' ) def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(F'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
37
0
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class _snake_case ( __a): def __init__( self : str, __lowercase : Dict=0.01, __lowercase : Optional[Any]=1000 ): lowercase__ = p_stop lowercase__ = max_length def __iter__( self : Tuple ): lowercase__ = 0 lowercase__ = False while not stop and count < self.max_length: yield count count += 1 lowercase__ = random.random() < self.p_stop class _snake_case ( unittest.TestCase): def A__ ( self : List[Any], __lowercase : List[Any], __lowercase : List[str], __lowercase : Optional[Any]=False, __lowercase : List[Any]=True ): lowercase__ = [ BatchSamplerShard(lowerCAmelCase_, 2, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) for i in range(2 ) ] lowercase__ = [list(lowerCAmelCase_ ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowerCAmelCase_ ) for shard in batch_sampler_shards], [len(lowerCAmelCase_ ) for e in expected] ) self.assertListEqual(lowerCAmelCase_, lowerCAmelCase_ ) def A__ ( self : int ): lowercase__ = BatchSampler(range(24 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) lowercase__ = BatchSampler(range(24 ), batch_size=3, drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowercase__ = BatchSampler(range(21 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) lowercase__ = BatchSampler(range(21 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowercase__ = BatchSampler(range(22 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) lowercase__ = BatchSampler(range(22 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowercase__ = BatchSampler(range(20 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) lowercase__ = BatchSampler(range(20 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) # Check the shards when the dataset is very small. lowercase__ = BatchSampler(range(2 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) lowercase__ = BatchSampler(range(2 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_ ) def A__ ( self : Any ): lowercase__ = BatchSampler(range(24 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(24 ), batch_size=4, drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowercase__ = BatchSampler(range(22 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(22 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowercase__ = BatchSampler(range(21 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(21 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. lowercase__ = BatchSampler(range(2 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(2 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_ ) def A__ ( self : Any ): lowercase__ = BatchSampler(range(24 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(24 ), batch_size=3, drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowercase__ = BatchSampler(range(21 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(21 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowercase__ = BatchSampler(range(22 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(22 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowercase__ = BatchSampler(range(20 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(20 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. lowercase__ = BatchSampler(range(2 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(2 ), batch_size=3, drop_last=lowerCAmelCase_ ) lowercase__ = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) def A__ ( self : Tuple ): lowercase__ = BatchSampler(range(24 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(24 ), batch_size=4, drop_last=lowerCAmelCase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowercase__ = BatchSampler(range(22 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(22 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowercase__ = BatchSampler(range(21 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(21 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) # Check the shards when the dataset is very small. lowercase__ = BatchSampler(range(2 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [[[0, 1]], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) lowercase__ = BatchSampler(range(2 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = [[], []] self.check_batch_sampler_shards(lowerCAmelCase_, lowerCAmelCase_, split_batches=lowerCAmelCase_, even_batches=lowerCAmelCase_ ) def A__ ( self : Optional[int] ): lowercase__ = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] lowercase__ = [BatchSamplerShard(lowerCAmelCase_, 2, lowerCAmelCase_, even_batches=lowerCAmelCase_ ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ), 3 ) self.assertEqual(len(batch_sampler_shards[1] ), 2 ) self.assertListEqual(list(batch_sampler_shards[0] ), [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ), [[3, 4], [9, 10, 11]] ) def A__ ( self : Optional[int], __lowercase : List[Any], __lowercase : Optional[int], __lowercase : str, __lowercase : str=False, __lowercase : List[str]=2, __lowercase : Optional[int]=False ): random.seed(lowerCAmelCase_ ) lowercase__ = list(lowerCAmelCase_ ) lowercase__ = [ IterableDatasetShard( lowerCAmelCase_, batch_size=lowerCAmelCase_, drop_last=lowerCAmelCase_, num_processes=lowerCAmelCase_, process_index=lowerCAmelCase_, split_batches=lowerCAmelCase_, ) for i in range(lowerCAmelCase_ ) ] lowercase__ = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowerCAmelCase_ ) iterable_dataset_lists.append(list(lowerCAmelCase_ ) ) lowercase__ = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size lowercase__ = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowerCAmelCase_ ), len(lowerCAmelCase_ ) ) self.assertTrue(len(lowerCAmelCase_ ) % shard_batch_size == 0 ) lowercase__ = [] for idx in range(0, len(lowerCAmelCase_ ), lowerCAmelCase_ ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowerCAmelCase_ ) < len(lowerCAmelCase_ ): reference += reference self.assertListEqual(lowerCAmelCase_, reference[: len(lowerCAmelCase_ )] ) def A__ ( self : str ): lowercase__ = 42 lowercase__ = RandomIterableDataset() self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) # Edge case with a very small dataset lowercase__ = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) self.check_iterable_dataset_shards(lowerCAmelCase_, lowerCAmelCase_, batch_size=4, drop_last=lowerCAmelCase_, split_batches=lowerCAmelCase_ ) def A__ ( self : Tuple ): lowercase__ = BatchSampler(range(16 ), batch_size=4, drop_last=lowerCAmelCase_ ) lowercase__ = SkipBatchSampler(lowerCAmelCase_, 2 ) self.assertListEqual(list(lowerCAmelCase_ ), [[8, 9, 10, 11], [12, 13, 14, 15]] ) def A__ ( self : Optional[int] ): lowercase__ = SkipDataLoader(list(range(16 ) ), batch_size=4, skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def A__ ( self : Union[str, Any] ): lowercase__ = DataLoader(list(range(16 ) ), batch_size=4 ) lowercase__ = skip_first_batches(lowerCAmelCase_, num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def A__ ( self : str ): lowercase__ = DataLoaderShard(list(range(16 ) ), batch_size=4 ) for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) def A__ ( self : Any ): Accelerator() lowercase__ = DataLoaderDispatcher(range(16 ), batch_size=4 ) for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowerCAmelCase_ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 )
718
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import math from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { 'facebook/data2vec-base-960h': 'https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json', # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class _snake_case ( __lowercase): UpperCamelCase__ : Union[str, Any] ='''data2vec-audio''' def __init__( self : Dict, __lowercase : Tuple=32, __lowercase : str=768, __lowercase : str=12, __lowercase : Tuple=12, __lowercase : Union[str, Any]=3072, __lowercase : Union[str, Any]="gelu", __lowercase : str=0.1, __lowercase : Optional[Any]=0.1, __lowercase : Any=0.1, __lowercase : str=0.0, __lowercase : Union[str, Any]=0.1, __lowercase : Optional[Any]=0.1, __lowercase : Optional[int]=0.02, __lowercase : Union[str, Any]=1e-5, __lowercase : Tuple="gelu", __lowercase : Tuple=(512, 512, 512, 512, 512, 512, 512), __lowercase : Union[str, Any]=(5, 2, 2, 2, 2, 2, 2), __lowercase : List[Any]=(10, 3, 3, 3, 3, 2, 2), __lowercase : Optional[Any]=False, __lowercase : Optional[Any]=16, __lowercase : Dict=19, __lowercase : Optional[Any]=5, __lowercase : List[str]=0.05, __lowercase : Union[str, Any]=10, __lowercase : Tuple=2, __lowercase : int=0.0, __lowercase : Union[str, Any]=10, __lowercase : str=0, __lowercase : Union[str, Any]="sum", __lowercase : Union[str, Any]=False, __lowercase : Dict=False, __lowercase : Optional[int]=256, __lowercase : Any=(512, 512, 512, 512, 1500), __lowercase : Optional[int]=(5, 3, 3, 1, 1), __lowercase : List[str]=(1, 2, 3, 1, 1), __lowercase : List[str]=512, __lowercase : Union[str, Any]=0, __lowercase : Dict=1, __lowercase : Union[str, Any]=2, __lowercase : Tuple=False, __lowercase : str=3, __lowercase : Dict=2, __lowercase : Optional[Any]=3, __lowercase : Optional[Any]=None, **__lowercase : str, ): super().__init__(**__a, pad_token_id=__a, bos_token_id=__a, eos_token_id=__a ) lowercase__ = hidden_size lowercase__ = feat_extract_activation lowercase__ = list(__a ) lowercase__ = list(__a ) lowercase__ = list(__a ) lowercase__ = conv_bias lowercase__ = num_conv_pos_embeddings lowercase__ = num_conv_pos_embedding_groups lowercase__ = conv_pos_kernel_size lowercase__ = len(self.conv_dim ) lowercase__ = num_hidden_layers lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = num_attention_heads lowercase__ = hidden_dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = feat_proj_dropout lowercase__ = final_dropout lowercase__ = layerdrop lowercase__ = layer_norm_eps lowercase__ = initializer_range lowercase__ = vocab_size lowercase__ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase__ = mask_time_prob lowercase__ = mask_time_length lowercase__ = mask_time_min_masks lowercase__ = mask_feature_prob lowercase__ = mask_feature_length lowercase__ = mask_feature_min_masks # ctc loss lowercase__ = ctc_loss_reduction lowercase__ = ctc_zero_infinity # adapter lowercase__ = add_adapter lowercase__ = adapter_kernel_size lowercase__ = adapter_stride lowercase__ = num_adapter_layers lowercase__ = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase__ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase__ = list(__a ) lowercase__ = list(__a ) lowercase__ = list(__a ) lowercase__ = xvector_output_dim @property def A__ ( self : Tuple ): return math.prod(self.conv_stride )
719
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowercase_ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : List[Any], __lowercase : int, __lowercase : Optional[int]=7, __lowercase : List[str]=3, __lowercase : Tuple=18, __lowercase : List[Any]=30, __lowercase : Tuple=400, __lowercase : Any=None, __lowercase : Optional[int]=True, __lowercase : List[str]=True, __lowercase : Union[str, Any]=None, ): lowercase__ = size if size is not None else {"height": 20, "width": 20} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = size lowercase__ = do_normalize lowercase__ = do_convert_rgb lowercase__ = [512, 1024, 2048, 4096] lowercase__ = patch_size if patch_size is not None else {"height": 16, "width": 16} def A__ ( self : List[str] ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def A__ ( self : Any ): lowercase__ = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" lowercase__ = Image.open(requests.get(__lowercase, stream=__lowercase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Any =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Any ): lowercase__ = PixaStructImageProcessingTester(self ) @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Optional[int] ): lowercase__ = self.image_processor_tester.prepare_dummy_image() lowercase__ = self.image_processing_class(**self.image_processor_dict ) lowercase__ = 2048 lowercase__ = image_processor(__lowercase, return_tensors="pt", max_patches=__lowercase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean(), torch.tensor(0.0606 ), atol=1e-3, rtol=1e-3 ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : int ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 lowercase__ = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__lowercase ): lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches lowercase__ = "Hello" lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Tuple ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Any ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Optional[int] ): lowercase__ = PixaStructImageProcessingTester(self, num_channels=4 ) lowercase__ = 3 @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), )
37
0
'''simple docstring''' import numpy as np def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return 1 / (1 + np.exp(-vector )) if __name__ == "__main__": import doctest doctest.testmod()
720
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = len(SCREAMING_SNAKE_CASE_ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ = 0 count += depth_first_search(SCREAMING_SNAKE_CASE_ , row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , row - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col - 1 , SCREAMING_SNAKE_CASE_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
37
0
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowercase_ = pd.read_csv("""sample_data.csv""", header=None) lowercase_ = df.shape[:1][0] # If you're using some other dataset input the target column lowercase_ = df.iloc[:, 1:2] lowercase_ = actual_data.values.reshape(len_data, 1) lowercase_ = MinMaxScaler().fit_transform(actual_data) lowercase_ = 10 lowercase_ = 5 lowercase_ = 20 lowercase_ = len_data - periods * look_back lowercase_ = actual_data[:division] lowercase_ = actual_data[division - look_back :] lowercase_ , lowercase_ = [], [] lowercase_ , lowercase_ = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowercase_ = np.array(train_x) lowercase_ = np.array(test_x) lowercase_ = np.array([list(i.ravel()) for i in train_y]) lowercase_ = np.array([list(i.ravel()) for i in test_y]) lowercase_ = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") lowercase_ = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) lowercase_ = model.predict(x_test)
721
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 for ch in input_str: lowercase__ = ord(SCREAMING_SNAKE_CASE_ ) lowercase__ = pow(2 , SCREAMING_SNAKE_CASE_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
37
0
'''simple docstring''' class _snake_case : def __init__( self : List[Any] ): lowercase__ = "" lowercase__ = "" lowercase__ = [] def A__ ( self : List[str], __lowercase : int, __lowercase : int ): if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: lowercase__ = self.__min_dist_top_down_dp(m - 1, n - 1 ) else: lowercase__ = self.__min_dist_top_down_dp(__lowercase, n - 1 ) lowercase__ = self.__min_dist_top_down_dp(m - 1, __lowercase ) lowercase__ = self.__min_dist_top_down_dp(m - 1, n - 1 ) lowercase__ = 1 + min(__lowercase, __lowercase, __lowercase ) return self.dp[m][n] def A__ ( self : Optional[int], __lowercase : str, __lowercase : str ): lowercase__ = worda lowercase__ = worda lowercase__ = [[-1 for _ in range(len(__lowercase ) )] for _ in range(len(__lowercase ) )] return self.__min_dist_top_down_dp(len(__lowercase ) - 1, len(__lowercase ) - 1 ) def A__ ( self : Optional[int], __lowercase : str, __lowercase : str ): lowercase__ = worda lowercase__ = worda lowercase__ = len(__lowercase ) lowercase__ = len(__lowercase ) lowercase__ = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty lowercase__ = j elif j == 0: # second string is empty lowercase__ = i elif worda[i - 1] == worda[j - 1]: # last characters are equal lowercase__ = self.dp[i - 1][j - 1] else: lowercase__ = self.dp[i][j - 1] lowercase__ = self.dp[i - 1][j] lowercase__ = self.dp[i - 1][j - 1] lowercase__ = 1 + min(__lowercase, __lowercase, __lowercase ) return self.dp[m][n] if __name__ == "__main__": lowercase_ = EditDistance() print("""****************** Testing Edit Distance DP Algorithm ******************""") print() lowercase_ = input("""Enter the first string: """).strip() lowercase_ = input("""Enter the second string: """).strip() print() print(F'The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}') print(F'The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}') print() print("""*************** End of Testing Edit Distance DP Algorithm ***************""")
700
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = len(SCREAMING_SNAKE_CASE_ ) for i in range(SCREAMING_SNAKE_CASE_ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE_ ): if numbers[j] < numbers[i]: lowercase__ , lowercase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() lowercase_ = [int(item) for item in user_input.split(""",""")] print(exchange_sort(unsorted))
37
0
from math import sqrt def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 for i in range(1 , int(sqrt(lowercase__ ) + 1 ) ): if n % i == 0 and i != sqrt(lowercase__ ): total += i + n // i elif i == sqrt(lowercase__ ): total += i return total - n def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = 1_0000 ): lowercase__ = sum( i for i in range(1 , lowercase__ ) if sum_of_divisors(sum_of_divisors(lowercase__ ) ) == i and sum_of_divisors(lowercase__ ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
701
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) lowercase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase__ = 1 if upper_limit > 0: lowercase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: lowercase_ = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F'The Catalan numbers from 0 through {N} are:') print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
37
0
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'spiece.model'} lowercase_ = { 'vocab_file': { 't5-small': 'https://huggingface.co/t5-small/resolve/main/spiece.model', 't5-base': 'https://huggingface.co/t5-base/resolve/main/spiece.model', 't5-large': 'https://huggingface.co/t5-large/resolve/main/spiece.model', 't5-3b': 'https://huggingface.co/t5-3b/resolve/main/spiece.model', 't5-11b': 'https://huggingface.co/t5-11b/resolve/main/spiece.model', } } # TODO(PVP) - this should be removed in Transformers v5 lowercase_ = { 't5-small': 512, 't5-base': 512, 't5-large': 512, 't5-3b': 512, 't5-11b': 512, } lowercase_ = '▁' class _snake_case ( a__): UpperCamelCase__ : Optional[int] =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Optional[Any] =["""input_ids""", """attention_mask"""] def __init__( self : List[str], __lowercase : int, __lowercase : Any="</s>", __lowercase : Tuple="<unk>", __lowercase : Optional[int]="<pad>", __lowercase : Optional[Any]=100, __lowercase : Union[str, Any]=None, __lowercase : List[Any] = None, __lowercase : Any=True, **__lowercase : Dict, ): if extra_ids > 0 and additional_special_tokens is None: lowercase__ = [F'''<extra_id_{i}>''' for i in range(lowercase__ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens lowercase__ = len(set(filter(lambda __lowercase : bool("extra_id" in str(lowercase__ ) ), lowercase__ ) ) ) if extra_tokens != extra_ids: raise ValueError( F'''Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are''' " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens" ) if legacy: logger.warning_once( F'''You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to''' " read the related pull request available at https://github.com/huggingface/transformers/pull/24565" ) lowercase__ = legacy lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase__, unk_token=lowercase__, pad_token=lowercase__, extra_ids=lowercase__, additional_special_tokens=lowercase__, sp_model_kwargs=self.sp_model_kwargs, legacy=lowercase__, **lowercase__, ) lowercase__ = vocab_file lowercase__ = extra_ids lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase__ ) @staticmethod def A__ ( __lowercase : int, __lowercase : str, __lowercase : Optional[int] ): if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: lowercase__ = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( "This tokenizer was incorrectly instantiated with a model max length of" F''' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this''' " behavior is kept to avoid breaking backwards compatibility when padding/encoding with" " `truncation is True`.\n- Be aware that you SHOULD NOT rely on" F''' {pretrained_model_name_or_path} automatically truncating your input to''' F''' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences''' F''' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with''' " `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please" " instantiate this tokenizer with `model_max_length` set to your preferred value.", lowercase__, ) return max_model_length @property def A__ ( self : Tuple ): return self.sp_model.get_piece_size() + self._extra_ids def A__ ( self : List[Any] ): lowercase__ = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A__ ( self : str, __lowercase : Tuple, __lowercase : Optional[int] = None, __lowercase : List[Any] = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase__, token_ids_a=lowercase__, already_has_special_tokens=lowercase__ ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowercase__ )) + [1] return ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] def A__ ( self : Optional[Any] ): return list( set(filter(lambda __lowercase : bool(re.search(R"<extra_id_\d+>", lowercase__ ) ) is not None, self.additional_special_tokens ) ) ) def A__ ( self : List[str] ): return [self._convert_token_to_id(lowercase__ ) for token in self.get_sentinel_tokens()] def A__ ( self : Dict, __lowercase : str ): if len(lowercase__ ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'''This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated''' " eos tokens being added." ) return token_ids else: return token_ids + [self.eos_token_id] def A__ ( self : Any, __lowercase : str, __lowercase : Union[str, Any] = None ): lowercase__ = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def A__ ( self : Optional[Any], __lowercase : List[Any], __lowercase : str = None ): lowercase__ = self._add_eos_if_not_present(lowercase__ ) if token_ids_a is None: return token_ids_a else: lowercase__ = self._add_eos_if_not_present(lowercase__ ) return token_ids_a + token_ids_a def __getstate__( self : str ): lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self : int, __lowercase : List[str] ): lowercase__ = d # for backward compatibility if not hasattr(self, "sp_model_kwargs" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A__ ( self : Tuple, __lowercase : Union[str, Any], **__lowercase : Optional[int] ): if not self.legacy: lowercase__ = SPIECE_UNDERLINE + text.replace(lowercase__, " " ) return super().tokenize(lowercase__, **lowercase__ ) def A__ ( self : Optional[Any], __lowercase : Tuple, **__lowercase : Dict ): if not self.legacy: lowercase__ = text.startswith(lowercase__ ) if is_first: lowercase__ = text[1:] lowercase__ = self.sp_model.encode(lowercase__, out_type=lowercase__ ) if not self.legacy and not is_first and not text.startswith(" " ) and tokens[0].startswith(lowercase__ ): lowercase__ = ([tokens[0][1:]] if len(tokens[0] ) > 1 else []) + tokens[1:] return tokens def A__ ( self : Union[str, Any], __lowercase : str ): if token.startswith("<extra_id_" ): lowercase__ = re.match(R"<extra_id_(\d+)>", lowercase__ ) lowercase__ = int(match.group(1 ) ) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(lowercase__ ) def A__ ( self : Optional[Any], __lowercase : Tuple ): if index < self.sp_model.get_piece_size(): lowercase__ = self.sp_model.IdToPiece(lowercase__ ) else: lowercase__ = F'''<extra_id_{self.vocab_size - 1 - index}>''' return token def A__ ( self : Dict, __lowercase : str ): lowercase__ = [] lowercase__ = '''''' lowercase__ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase__ ) + token lowercase__ = True lowercase__ = [] else: current_sub_tokens.append(lowercase__ ) lowercase__ = False out_string += self.sp_model.decode(lowercase__ ) return out_string.strip() def A__ ( self : Optional[int], __lowercase : str, __lowercase : Union[str, Any] = None ): if not os.path.isdir(lowercase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( lowercase__, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, lowercase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowercase__, "wb" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(lowercase__ ) return (out_vocab_file,)
702
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase_ = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json""" ), } class _snake_case ( _UpperCAmelCase): UpperCamelCase__ ="""dpr""" def __init__( self : int, __lowercase : Any=3_0522, __lowercase : int=768, __lowercase : Optional[int]=12, __lowercase : Optional[Any]=12, __lowercase : Any=3072, __lowercase : Union[str, Any]="gelu", __lowercase : Dict=0.1, __lowercase : List[Any]=0.1, __lowercase : Union[str, Any]=512, __lowercase : str=2, __lowercase : Dict=0.02, __lowercase : Union[str, Any]=1e-1_2, __lowercase : Union[str, Any]=0, __lowercase : List[Any]="absolute", __lowercase : str = 0, **__lowercase : str, ): super().__init__(pad_token_id=lowercase__, **lowercase__ ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = projection_dim lowercase__ = position_embedding_type
703
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") lowercase__ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f'''bert/{name}''' def create_tf_var(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = session.run(SCREAMING_SNAKE_CASE_ ) print(f'''Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace("-" , "_" ) + ".ckpt" ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_=None ): lowercase__ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory in which to save tensorflow model" ) lowercase__ = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
37
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase_ = {"""LayoutLMv2Config""", """LayoutLMv3Config"""} @is_pipeline_test class _snake_case ( unittest.TestCase): UpperCamelCase__ : Dict =MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase__ : int =TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase__ : Optional[int] ={config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase__ : Optional[Any] ={ config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def A__ ( self : Dict, __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : Dict ): lowercase__ = ZeroShotClassificationPipeline( model=__snake_case, tokenizer=__snake_case, candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def A__ ( self : int, __lowercase : str, __lowercase : List[Any] ): lowercase__ = classifier("Who are you voting for in 2020?", candidate_labels="politics" ) self.assertEqual(__snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case )], "scores": [ANY(__snake_case )]} ) # No kwarg lowercase__ = classifier("Who are you voting for in 2020?", ["politics"] ) self.assertEqual(__snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case )], "scores": [ANY(__snake_case )]} ) lowercase__ = classifier("Who are you voting for in 2020?", candidate_labels=["politics"] ) self.assertEqual(__snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case )], "scores": [ANY(__snake_case )]} ) lowercase__ = classifier("Who are you voting for in 2020?", candidate_labels="politics, public health" ) self.assertEqual( __snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case ), ANY(__snake_case )], "scores": [ANY(__snake_case ), ANY(__snake_case )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ), 1.0 ) lowercase__ = classifier("Who are you voting for in 2020?", candidate_labels=["politics", "public health"] ) self.assertEqual( __snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case ), ANY(__snake_case )], "scores": [ANY(__snake_case ), ANY(__snake_case )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ), 1.0 ) lowercase__ = classifier( "Who are you voting for in 2020?", candidate_labels="politics", hypothesis_template="This text is about {}" ) self.assertEqual(__snake_case, {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case )], "scores": [ANY(__snake_case )]} ) # https://github.com/huggingface/transformers/issues/13846 lowercase__ = classifier(["I am happy"], ["positive", "negative"] ) self.assertEqual( __snake_case, [ {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case ), ANY(__snake_case )], "scores": [ANY(__snake_case ), ANY(__snake_case )]} for i in range(1 ) ], ) lowercase__ = classifier(["I am happy", "I am sad"], ["positive", "negative"] ) self.assertEqual( __snake_case, [ {"sequence": ANY(__snake_case ), "labels": [ANY(__snake_case ), ANY(__snake_case )], "scores": [ANY(__snake_case ), ANY(__snake_case )]} for i in range(2 ) ], ) with self.assertRaises(__snake_case ): classifier("", candidate_labels="politics" ) with self.assertRaises(__snake_case ): classifier(__snake_case, candidate_labels="politics" ) with self.assertRaises(__snake_case ): classifier("Who are you voting for in 2020?", candidate_labels="" ) with self.assertRaises(__snake_case ): classifier("Who are you voting for in 2020?", candidate_labels=__snake_case ) with self.assertRaises(__snake_case ): classifier( "Who are you voting for in 2020?", candidate_labels="politics", hypothesis_template="Not formatting template", ) with self.assertRaises(__snake_case ): classifier( "Who are you voting for in 2020?", candidate_labels="politics", hypothesis_template=__snake_case, ) self.run_entailment_id(__snake_case ) def A__ ( self : Dict, __lowercase : Pipeline ): lowercase__ = zero_shot_classifier.model.config lowercase__ = config.labelaid lowercase__ = zero_shot_classifier.entailment_id lowercase__ = {'''LABEL_0''': 0, '''LABEL_1''': 1, '''LABEL_2''': 2} self.assertEqual(zero_shot_classifier.entailment_id, -1 ) lowercase__ = {'''entailment''': 0, '''neutral''': 1, '''contradiction''': 2} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) lowercase__ = {'''ENTAIL''': 0, '''NON-ENTAIL''': 1} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) lowercase__ = {'''ENTAIL''': 2, '''NEUTRAL''': 1, '''CONTR''': 0} self.assertEqual(zero_shot_classifier.entailment_id, 2 ) lowercase__ = original_labelaid self.assertEqual(__snake_case, zero_shot_classifier.entailment_id ) @require_torch def A__ ( self : str ): lowercase__ = pipeline( "zero-shot-classification", model="sshleifer/tiny-distilbert-base-cased-distilled-squad", framework="pt", ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100, candidate_labels=["politics", "public health", "science"] ) @require_torch def A__ ( self : Optional[Any] ): lowercase__ = pipeline( "zero-shot-classification", model="sshleifer/tiny-distilbert-base-cased-distilled-squad", framework="pt", ) lowercase__ = zero_shot_classifier( "Who are you voting for in 2020?", candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.333, 0.333, 0.333], }, ) @require_tf def A__ ( self : Union[str, Any] ): lowercase__ = pipeline( "zero-shot-classification", model="sshleifer/tiny-distilbert-base-cased-distilled-squad", framework="tf", ) lowercase__ = zero_shot_classifier( "Who are you voting for in 2020?", candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.333, 0.333, 0.333], }, ) @slow @require_torch def A__ ( self : Dict ): lowercase__ = pipeline("zero-shot-classification", model="roberta-large-mnli", framework="pt" ) lowercase__ = zero_shot_classifier( "Who are you voting for in 2020?", candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.976, 0.015, 0.009], }, ) lowercase__ = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data.", candidate_labels=["machine learning", "statistics", "translation", "vision"], multi_label=__snake_case, ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.817, 0.713, 0.018, 0.018], }, ) @slow @require_tf def A__ ( self : Tuple ): lowercase__ = pipeline("zero-shot-classification", model="roberta-large-mnli", framework="tf" ) lowercase__ = zero_shot_classifier( "Who are you voting for in 2020?", candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.976, 0.015, 0.009], }, ) lowercase__ = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data.", candidate_labels=["machine learning", "statistics", "translation", "vision"], multi_label=__snake_case, ) self.assertEqual( nested_simplify(__snake_case ), { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.817, 0.713, 0.018, 0.018], }, )
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_timesformer""": ["""TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimesformerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimesformerModel""", """TimesformerForVideoClassification""", """TimesformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [0 for i in range(r + 1 )] # nc0 = 1 lowercase__ = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. lowercase__ = min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=10, r=5))
705
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() lowercase_ = { """bart""": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), """bert""": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-base-cased-finetuned-mrpc""": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """dpr""": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), """gpt2""": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlnet""": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm""": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm-roberta""": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """transfo-xl""": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """openai-gpt""": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """roberta""": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """layoutlm""": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), """roberta-large-mnli""": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """camembert""": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """flaubert""": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert""": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert-base-distilled-squad""": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert""": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert-visual-feature-encoder""": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """ctrl""": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """albert""": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """t5""": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """electra""": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """wav2vec2""": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) lowercase__ = config_class.from_json_file(SCREAMING_SNAKE_CASE_ ) lowercase__ = True lowercase__ = True print(f'''Building TensorFlow model from configuration: {config}''' ) lowercase__ = model_class(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): lowercase__ = cached_file( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: lowercase__ = load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if compare_with_pt_model: lowercase__ = tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE_ ) # build the network lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" ) lowercase__ = pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE_ , config=SCREAMING_SNAKE_CASE_ , state_dict=SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): lowercase__ = pt_model(**pt_model.dummy_inputs ) lowercase__ = pto[0].numpy() lowercase__ = tfo[0].numpy() lowercase__ = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(SCREAMING_SNAKE_CASE_ , save_format="h5" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ): if args_model_type is None: lowercase__ = list(MODEL_CLASSES.keys() ) else: lowercase__ = [args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE_ , start=1 ): print("=" * 100 ) print(f''' Converting model type {j}/{len(SCREAMING_SNAKE_CASE_ )}: {model_type}''' ) print("=" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: lowercase__ = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: lowercase__ = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , start=1 ): print("-" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue lowercase__ = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE_ )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 100 ) if config_shortcut_name in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = config_shortcut_name if model_shortcut_name in aws_model_maps: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE_ ): lowercase__ = "converted_model" convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE_ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE_ , config_file=SCREAMING_SNAKE_CASE_ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE_ , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=SCREAMING_SNAKE_CASE_ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE_ ) os.remove(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_dump_path""", default=None, type=str, required=True, help="""Path to the output Tensorflow dump file.""" ) parser.add_argument( """--model_type""", default=None, type=str, help=( F'Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ' """convert all the models from AWS.""" ), ) parser.add_argument( """--pytorch_checkpoint_path""", default=None, type=str, help=( """Path to the PyTorch checkpoint path or shortcut name to download from AWS. """ """If not given, will download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--config_file""", default=None, type=str, help=( """The config json file corresponding to the pre-trained model. \n""" """This specifies the model architecture. If not given and """ """--pytorch_checkpoint_path is not given or is a shortcut name """ """use the configuration associated to the shortcut name on the AWS""" ), ) parser.add_argument( """--compare_with_pt_model""", action="""store_true""", help="""Compare Tensorflow and PyTorch model predictions.""" ) parser.add_argument( """--use_cached_models""", action="""store_true""", help="""Use cached models if possible instead of updating to latest checkpoint versions.""", ) parser.add_argument( """--remove_cached_files""", action="""store_true""", help="""Remove pytorch models after conversion (save memory when converting in batches).""", ) parser.add_argument("""--only_convert_finetuned_models""", action="""store_true""", help="""Only convert finetuned models.""") lowercase_ = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
37
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType lowercase_ = logging.get_logger(__name__) lowercase_ = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off lowercase_ = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 1_0563, 1_0786, 1_1420, 1_1709, 1_1907, 1_3163, 1_3697, 1_3700, 1_4808, 1_5306, 1_6410, 1_6791, 1_7992, 1_9203, 1_9510, 2_0724, 2_2305, 2_2935, 2_7007, 3_0109, 3_0420, 3_3409, 3_4949, 4_0283, 4_0493, 4_0549, 4_7282, 4_9146, 5_0257, 5_0359, 5_0360, 5_0361 ] lowercase_ = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 1_0428, 1_0929, 1_1938, 1_2033, 1_2331, 1_2562, 1_3793, 1_4157, 1_4635, 1_5265, 1_5618, 1_6553, 1_6604, 1_8362, 1_8956, 2_0075, 2_1675, 2_2520, 2_6130, 2_6161, 2_6435, 2_8279, 2_9464, 3_1650, 3_2302, 3_2470, 3_6865, 4_2863, 4_7425, 4_9870, 5_0254, 5_0258, 5_0360, 5_0361, 5_0362 ] class _snake_case ( _UpperCamelCase): UpperCamelCase__ : List[Any] ="whisper" UpperCamelCase__ : Dict =["past_key_values"] UpperCamelCase__ : List[Any] ={"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any], __lowercase : Dict=5_1865, __lowercase : Tuple=80, __lowercase : Optional[Any]=6, __lowercase : List[Any]=4, __lowercase : Any=6, __lowercase : Tuple=4, __lowercase : List[str]=1536, __lowercase : str=1536, __lowercase : Any=0.0, __lowercase : List[str]=0.0, __lowercase : Dict=5_0257, __lowercase : str=True, __lowercase : Optional[Any]=True, __lowercase : Optional[Any]="gelu", __lowercase : Optional[int]=256, __lowercase : List[Any]=0.0, __lowercase : Dict=0.0, __lowercase : List[Any]=0.0, __lowercase : int=0.02, __lowercase : Optional[Any]=False, __lowercase : Optional[Any]=1500, __lowercase : List[str]=448, __lowercase : Union[str, Any]=5_0256, __lowercase : List[str]=5_0256, __lowercase : Any=5_0256, __lowercase : str=None, __lowercase : List[Any]=[220, 5_0256], __lowercase : int=False, __lowercase : str=256, __lowercase : Any=False, __lowercase : Optional[Any]=0.05, __lowercase : Union[str, Any]=10, __lowercase : int=2, __lowercase : int=0.0, __lowercase : Optional[int]=10, __lowercase : str=0, __lowercase : Any=7, **__lowercase : Tuple, ): lowercase__ = vocab_size lowercase__ = num_mel_bins lowercase__ = d_model lowercase__ = encoder_layers lowercase__ = encoder_attention_heads lowercase__ = decoder_layers lowercase__ = decoder_attention_heads lowercase__ = decoder_ffn_dim lowercase__ = encoder_ffn_dim lowercase__ = dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = activation_function lowercase__ = init_std lowercase__ = encoder_layerdrop lowercase__ = decoder_layerdrop lowercase__ = use_cache lowercase__ = encoder_layers lowercase__ = scale_embedding # scale factor will be sqrt(d_model) if True lowercase__ = max_source_positions lowercase__ = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. lowercase__ = classifier_proj_size lowercase__ = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase__ = apply_spec_augment lowercase__ = mask_time_prob lowercase__ = mask_time_length lowercase__ = mask_time_min_masks lowercase__ = mask_feature_prob lowercase__ = mask_feature_length lowercase__ = mask_feature_min_masks lowercase__ = median_filter_width super().__init__( pad_token_id=__a, bos_token_id=__a, eos_token_id=__a, is_encoder_decoder=__a, decoder_start_token_id=__a, suppress_tokens=__a, begin_suppress_tokens=__a, **__a, ) class _snake_case ( _UpperCamelCase): @property def A__ ( self : Optional[int] ): lowercase__ = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ] ) if self.use_past: lowercase__ = {0: "batch"} else: lowercase__ = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__a, direction="inputs" ) return common_inputs def A__ ( self : Dict, __lowercase : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"], __lowercase : int = -1, __lowercase : int = -1, __lowercase : bool = False, __lowercase : Optional["TensorType"] = None, __lowercase : int = 2_2050, __lowercase : float = 5.0, __lowercase : int = 220, ): lowercase__ = OrderedDict() lowercase__ = OnnxConfig.generate_dummy_inputs( self, preprocessor=preprocessor.feature_extractor, batch_size=__a, framework=__a, sampling_rate=__a, time_duration=__a, frequency=__a, ) lowercase__ = encoder_inputs["input_features"].shape[2] lowercase__ = encoder_sequence_length // 2 if self.use_past else seq_length lowercase__ = super().generate_dummy_inputs( preprocessor.tokenizer, __a, __a, __a, __a ) lowercase__ = encoder_inputs.pop("input_features" ) lowercase__ = decoder_inputs.pop("decoder_input_ids" ) if "past_key_values" in decoder_inputs: lowercase__ = decoder_inputs.pop("past_key_values" ) return dummy_inputs @property def A__ ( self : str ): return 1e-3
706
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/config.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/config.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json""" ), """distilbert-base-uncased-finetuned-sst-2-english""": ( """https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json""" ), } class _snake_case ( UpperCAmelCase_): UpperCamelCase__ : int ="distilbert" UpperCamelCase__ : Tuple ={ "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self : Optional[Any], __lowercase : Tuple=3_0522, __lowercase : int=512, __lowercase : Optional[Any]=False, __lowercase : List[Any]=6, __lowercase : Any=12, __lowercase : Tuple=768, __lowercase : Tuple=4 * 768, __lowercase : int=0.1, __lowercase : Dict=0.1, __lowercase : str="gelu", __lowercase : List[Any]=0.02, __lowercase : Union[str, Any]=0.1, __lowercase : str=0.2, __lowercase : Union[str, Any]=0, **__lowercase : List[Any], ): lowercase__ = vocab_size lowercase__ = max_position_embeddings lowercase__ = sinusoidal_pos_embds lowercase__ = n_layers lowercase__ = n_heads lowercase__ = dim lowercase__ = hidden_dim lowercase__ = dropout lowercase__ = attention_dropout lowercase__ = activation lowercase__ = initializer_range lowercase__ = qa_dropout lowercase__ = seq_classif_dropout super().__init__(**_snake_case, pad_token_id=_snake_case ) class _snake_case ( UpperCAmelCase_): @property def A__ ( self : Optional[int] ): if self.task == "multiple-choice": lowercase__ = {0: "batch", 1: "choice", 2: "sequence"} else: lowercase__ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
707
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : def __init__( self : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Union[str, Any], __lowercase : str, __lowercase : List[Any], __lowercase : List[str]=0.2, __lowercase : List[str]=0.2 ): lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = conva_get[:2] lowercase__ = conva_get[2] lowercase__ = size_pa lowercase__ = rate_w lowercase__ = rate_t lowercase__ = [ np.mat(-1 * np.random.rand(self.conva[0], self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self : Any, __lowercase : List[str] ): # save model dict with pickle lowercase__ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(__lowercase, "wb" ) as f: pickle.dump(__lowercase, __lowercase ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls : Dict, __lowercase : Union[str, Any] ): # read saved model with open(__lowercase, "rb" ) as f: lowercase__ = pickle.load(__lowercase ) # noqa: S301 lowercase__ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) lowercase__ = model_dic.get("size_pooling1" ) lowercase__ = model_dic.get("num_bp1" ) lowercase__ = model_dic.get("num_bp2" ) lowercase__ = model_dic.get("num_bp3" ) lowercase__ = model_dic.get("rate_weight" ) lowercase__ = model_dic.get("rate_thre" ) # create model instance lowercase__ = CNN(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) # modify model parameter lowercase__ = model_dic.get("w_conv1" ) lowercase__ = model_dic.get("wkj" ) lowercase__ = model_dic.get("vji" ) lowercase__ = model_dic.get("thre_conv1" ) lowercase__ = model_dic.get("thre_bp2" ) lowercase__ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self : str, __lowercase : List[Any] ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self : List[str], __lowercase : Optional[Any] ): return round(__lowercase, 3 ) def A__ ( self : Optional[Any], __lowercase : Dict, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : str ): # convolution process lowercase__ = convs[0] lowercase__ = convs[1] lowercase__ = np.shape(__lowercase )[0] # get the data slice of original image data, data_focus lowercase__ = [] for i_focus in range(0, size_data - size_conv + 1, __lowercase ): for j_focus in range(0, size_data - size_conv + 1, __lowercase ): lowercase__ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ = [] lowercase__ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__lowercase ): lowercase__ = [] for i_focus in range(len(__lowercase ) ): lowercase__ = ( np.sum(np.multiply(data_focus[i_focus], w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape( __lowercase, __lowercase ) data_featuremap.append(__lowercase ) # expanding the data slice to One dimenssion lowercase__ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__lowercase ) ) lowercase__ = np.asarray(__lowercase ) return focus_list, data_featuremap def A__ ( self : List[Any], __lowercase : Any, __lowercase : List[Any], __lowercase : Union[str, Any]="average_pool" ): # pooling process lowercase__ = len(featuremaps[0] ) lowercase__ = int(size_map / size_pooling ) lowercase__ = [] for i_map in range(len(__lowercase ) ): lowercase__ = featuremaps[i_map] lowercase__ = [] for i_focus in range(0, __lowercase, __lowercase ): for j_focus in range(0, __lowercase, __lowercase ): lowercase__ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape(__lowercase, __lowercase ) featuremap_pooled.append(__lowercase ) return featuremap_pooled def A__ ( self : str, __lowercase : Optional[Any] ): # expanding three dimension data to one dimension list lowercase__ = [] for i in range(len(__lowercase ) ): lowercase__ = np.shape(data[i] ) lowercase__ = data[i].reshape(1, shapes[0] * shapes[1] ) lowercase__ = data_listed.getA().tolist()[0] data_expanded.extend(__lowercase ) lowercase__ = np.asarray(__lowercase ) return data_expanded def A__ ( self : Optional[int], __lowercase : Optional[int] ): # expanding matrix to one dimension list lowercase__ = np.asarray(__lowercase ) lowercase__ = np.shape(__lowercase ) lowercase__ = data_mat.reshape(1, shapes[0] * shapes[1] ) return data_expanded def A__ ( self : str, __lowercase : Tuple, __lowercase : List[Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = [] lowercase__ = 0 for i_map in range(__lowercase ): lowercase__ = np.ones((size_map, size_map) ) for i in range(0, __lowercase, __lowercase ): for j in range(0, __lowercase, __lowercase ): lowercase__ = pd_pool[ i_pool ] lowercase__ = i_pool + 1 lowercase__ = np.multiply( __lowercase, np.multiply(out_map[i_map], (1 - out_map[i_map]) ) ) pd_all.append(__lowercase ) return pd_all def A__ ( self : Tuple, __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : List[str]=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(__lowercase )) ) print((" - - Shape: Teach_Data ", np.shape(__lowercase )) ) lowercase__ = 0 lowercase__ = [] lowercase__ = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase__ = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(__lowercase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ = np.asmatrix(datas_train[p] ) lowercase__ = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = np.shape(__lowercase ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = np.dot(__lowercase, self.vji.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = np.dot(__lowercase, self.wkj.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ = np.multiply( (data_teach - bp_outa), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.multiply( np.dot(__lowercase, self.wkj ), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.dot(__lowercase, self.vji ) lowercase__ = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ = pd_conva_pooled.T.getA().tolist() lowercase__ = self._calculate_gradient_from_pool( __lowercase, __lowercase, shape_featuremapa[0], shape_featuremapa[1], self.size_poolinga, ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ = self.rate_weight * np.dot(__lowercase, __lowercase ) lowercase__ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ = rp + 1 lowercase__ = error_count / patterns all_mse.append(__lowercase ) def draw_error(): lowercase__ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__lowercase, "+-" ) plt.plot(__lowercase, "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(__lowercase, alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self : List[str], __lowercase : Optional[int] ): # model predict lowercase__ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(__lowercase )) ) for p in range(len(__lowercase ) ): lowercase__ = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = bp_outa * self.vji.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = bp_outa * self.wkj.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ = [list(map(self.do_round, __lowercase ) ) for each in produce_out] return np.asarray(__lowercase ) def A__ ( self : int, __lowercase : Any ): # return the data of image after convoluting process so we can check it out lowercase__ = np.asmatrix(__lowercase ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
37
0
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowercase_ = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class _snake_case ( _lowerCAmelCase , unittest.TestCase): UpperCamelCase__ : Any =DebertaVaTokenizer UpperCamelCase__ : Optional[Any] =DebertaVaTokenizerFast UpperCamelCase__ : int =True UpperCamelCase__ : str =True def A__ ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, unk_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self : str, __lowercase : List[str] ): lowercase__ = "this is a test" lowercase__ = "this is a test" return input_text, output_text def A__ ( self : Tuple ): lowercase__ = "<pad>" lowercase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ), _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ), _lowerCAmelCase ) def A__ ( self : Union[str, Any] ): lowercase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], "<pad>" ) self.assertEqual(vocab_keys[1], "<unk>" ) self.assertEqual(vocab_keys[-1], "[PAD]" ) self.assertEqual(len(_lowerCAmelCase ), 3_0001 ) def A__ ( self : Union[str, Any] ): self.assertEqual(self.get_tokenizer().vocab_size, 3_0000 ) def A__ ( self : int ): lowercase__ = " \tHeLLo!how \n Are yoU? " lowercase__ = ["▁hello", "!", "how", "▁are", "▁you", "?"] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, do_lower_case=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, do_lower_case=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) @unittest.skip("There is an inconsistency between slow and fast tokenizer due to a bug in the fast one." ) def A__ ( self : Union[str, Any] ): pass @unittest.skip("There is an inconsistency between slow and fast tokenizer due to a bug in the fast one." ) def A__ ( self : int ): pass def A__ ( self : str ): lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = ["▁", "<unk>", "▁was", "▁born", "▁in", "▁9", "2000", "▁", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "▁", ".", ] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : List[str] ): lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", "▁", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "▁", ".", ] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : int ): lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", ".", ] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : Tuple ): lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = ["▁", "<unk>", "▁was", "▁born", "▁in", "▁9", "2000", "▁", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "▁", ".", ] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : Tuple ): lowercase__ = " \tHeLLo!how \n Are yoU? " lowercase__ = ["▁", "<unk>", "e", "<unk>", "o", "!", "how", "▁", "<unk>", "re", "▁yo", "<unk>", "?"] # fmt: on lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, do_lower_case=_lowerCAmelCase, split_by_punct=_lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : Optional[int] ): lowercase__ = self.get_tokenizer() lowercase__ = self.get_rust_tokenizer() lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = tokenizer.convert_ids_to_tokens(tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) lowercase__ = rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = self.get_rust_tokenizer() lowercase__ = tokenizer.encode(_lowerCAmelCase ) lowercase__ = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : Dict ): lowercase__ = "This is a test" lowercase__ = [13, 1, 4398, 25, 21, 1289] lowercase__ = ["▁", "T", "his", "▁is", "▁a", "▁test"] lowercase__ = ["▁", "<unk>", "his", "▁is", "▁a", "▁test"] lowercase__ = DebertaVaTokenizer(_lowerCAmelCase, keep_accents=_lowerCAmelCase ) lowercase__ = DebertaVaTokenizerFast(_lowerCAmelCase, keep_accents=_lowerCAmelCase ) lowercase__ = tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) # fmt: off lowercase__ = "I was born in 92000, and this is falsé." lowercase__ = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] lowercase__ = ["▁", "I", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", ".", ] lowercase__ = ["▁", "<unk>", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", ".", ] # fmt: on lowercase__ = tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.encode(_lowerCAmelCase, add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) lowercase__ = rust_tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase, _lowerCAmelCase ) def A__ ( self : Union[str, Any] ): lowercase__ = DebertaVaTokenizer(_lowerCAmelCase ) lowercase__ = tokenizer.encode("sequence builders" ) lowercase__ = tokenizer.encode("multi-sequence build" ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase, _lowerCAmelCase ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id], _lowerCAmelCase ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id], _lowerCAmelCase, ) @slow def A__ ( self : Optional[Any] ): lowercase__ = {"input_ids": [[1, 3_9867, 36, 1_9390, 486, 27, 3_5052, 8_1436, 18, 6_0685, 1225, 7, 3_5052, 8_1436, 18, 9367, 1_6899, 18, 1_5937, 53, 594, 773, 18, 1_6287, 3_0465, 36, 1_5937, 6, 4_1139, 38, 3_6979, 6_0763, 191, 6, 3_4132, 99, 6, 5_0538, 390, 4_3230, 6, 3_4132, 2779, 2_0850, 14, 699, 1072, 1194, 36, 382, 1_0901, 53, 7, 699, 1072, 2084, 36, 2_0422, 630, 53, 19, 105, 3049, 1896, 1053, 1_6899, 1506, 11, 3_7978, 4243, 7, 1237, 3_1869, 200, 1_6566, 654, 6, 3_5052, 8_1436, 7, 5_5630, 1_3593, 4, 2], [1, 26, 1_5011, 13, 667, 8, 1053, 18, 2_3611, 1237, 7_2356, 1_2820, 34, 10_4134, 1209, 35, 1_3313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 1_5785, 1_4951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase, model_name="microsoft/deberta-v2-xlarge", revision="ad6e42c1532ddf3a15c39246b63f5559d558b670", )
708
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=SCREAMING_SNAKE_CASE_ , num_labels=1000 , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): lowercase__ = get_config(SCREAMING_SNAKE_CASE_ ) # load original model from timm lowercase__ = create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=SCREAMING_SNAKE_CASE_ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) lowercase__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # verify logits with torch.no_grad(): lowercase__ = model(SCREAMING_SNAKE_CASE_ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(SCREAMING_SNAKE_CASE_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE_ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowercase_ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
37
0
import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class _snake_case ( UpperCamelCase__ , unittest.TestCase): UpperCamelCase__ : Optional[Any] =PriorTransformer UpperCamelCase__ : Tuple ="""hidden_states""" @property def A__ ( self : List[Any] ): lowercase__ = 4 lowercase__ = 8 lowercase__ = 7 lowercase__ = floats_tensor((batch_size, embedding_dim) ).to(_a ) lowercase__ = floats_tensor((batch_size, embedding_dim) ).to(_a ) lowercase__ = floats_tensor((batch_size, num_embeddings, embedding_dim) ).to(_a ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def A__ ( self : List[str], __lowercase : List[Any]=0 ): torch.manual_seed(_a ) lowercase__ = 4 lowercase__ = 8 lowercase__ = 7 lowercase__ = torch.randn((batch_size, embedding_dim) ).to(_a ) lowercase__ = torch.randn((batch_size, embedding_dim) ).to(_a ) lowercase__ = torch.randn((batch_size, num_embeddings, embedding_dim) ).to(_a ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def A__ ( self : Tuple ): return (4, 8) @property def A__ ( self : Optional[int] ): return (4, 8) def A__ ( self : Optional[Any] ): lowercase__ = { """num_attention_heads""": 2, """attention_head_dim""": 4, """num_layers""": 2, """embedding_dim""": 8, """num_embeddings""": 7, """additional_embeddings""": 4, } lowercase__ = self.dummy_input return init_dict, inputs_dict def A__ ( self : int ): lowercase__ = PriorTransformer.from_pretrained( "hf-internal-testing/prior-dummy", output_loading_info=_a ) self.assertIsNotNone(_a ) self.assertEqual(len(loading_info["missing_keys"] ), 0 ) model.to(_a ) lowercase__ = model(**self.dummy_input )[0] assert hidden_states is not None, "Make sure output is not None" def A__ ( self : Optional[Any] ): lowercase__ = self.prepare_init_args_and_inputs_for_common() lowercase__ = self.model_class(**_a ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["""hidden_states""", """timestep"""] self.assertListEqual(arg_names[:2], _a ) def A__ ( self : Dict ): lowercase__ = PriorTransformer.from_pretrained("hf-internal-testing/prior-dummy" ) lowercase__ = model.to(_a ) if hasattr(_a, "set_default_attn_processor" ): model.set_default_attn_processor() lowercase__ = self.get_dummy_seed_input() with torch.no_grad(): lowercase__ = model(**_a )[0] lowercase__ = output[0, :5].flatten().cpu() print(_a ) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. lowercase__ = torch.tensor([-1.3436, -0.2870, 0.7538, 0.4368, -0.0239] ) self.assertTrue(torch_all_close(_a, _a, rtol=1e-2 ) ) @slow class _snake_case ( unittest.TestCase): def A__ ( self : Tuple, __lowercase : List[Any]=1, __lowercase : Optional[Any]=768, __lowercase : Optional[int]=77, __lowercase : List[str]=0 ): torch.manual_seed(_a ) lowercase__ = batch_size lowercase__ = embedding_dim lowercase__ = num_embeddings lowercase__ = torch.randn((batch_size, embedding_dim) ).to(_a ) lowercase__ = torch.randn((batch_size, embedding_dim) ).to(_a ) lowercase__ = torch.randn((batch_size, num_embeddings, embedding_dim) ).to(_a ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def A__ ( self : Tuple ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [13, [-0.5861, 0.1283, -0.0931, 0.0882, 0.4476, 0.1329, -0.0498, 0.0640]], [37, [-0.4913, 0.0110, -0.0483, 0.0541, 0.4954, -0.0170, 0.0354, 0.1651]], # fmt: on ] ) def A__ ( self : Tuple, __lowercase : Any, __lowercase : Optional[int] ): lowercase__ = PriorTransformer.from_pretrained("kandinsky-community/kandinsky-2-1-prior", subfolder="prior" ) model.to(_a ) lowercase__ = self.get_dummy_seed_input(seed=_a ) with torch.no_grad(): lowercase__ = model(**_a )[0] assert list(sample.shape ) == [1, 768] lowercase__ = sample[0, :8].flatten().cpu() print(_a ) lowercase__ = torch.tensor(_a ) assert torch_all_close(_a, _a, atol=1e-3 )
709
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _snake_case ( lowercase__): def __init__( self : Optional[Any], __lowercase : str = "▁", __lowercase : bool = True, __lowercase : Union[str, AddedToken] = "<unk>", __lowercase : Union[str, AddedToken] = "</s>", __lowercase : Union[str, AddedToken] = "<pad>", ): lowercase__ = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } lowercase__ = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowercase__ = token_dict["token"] lowercase__ = Tokenizer(Unigram() ) lowercase__ = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ), " " ), normalizers.Lowercase(), ] ) lowercase__ = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) lowercase__ = decoders.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ) lowercase__ = TemplateProcessing( single=F'''$A {self.special_tokens["eos"]["token"]}''', special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])], ) lowercase__ = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase, __lowercase ) def A__ ( self : Union[str, Any], __lowercase : Union[str, List[str]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) if isinstance(__lowercase, __lowercase ): lowercase__ = [files] self._tokenizer.train(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : List[Any], __lowercase : Union[Iterator[str], Iterator[Iterator[str]]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) self._tokenizer.train_from_iterator(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : str ): lowercase__ = json.loads(self._tokenizer.to_str() ) lowercase__ = self.special_tokens["unk"]["id"] lowercase__ = Tokenizer.from_str(json.dumps(__lowercase ) )
37
0
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class _snake_case ( unittest.TestCase): def A__ ( self : Optional[Any] ): lowercase__ = tempfile.mkdtemp() lowercase__ = SamImageProcessor() lowercase__ = SamProcessor(UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def A__ ( self : Optional[Any], **__lowercase : Optional[int] ): return AutoProcessor.from_pretrained(self.tmpdirname, **UpperCamelCase__ ).image_processor def A__ ( self : Union[str, Any] ): shutil.rmtree(self.tmpdirname ) def A__ ( self : str ): lowercase__ = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] lowercase__ = [Image.fromarray(np.moveaxis(UpperCamelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def A__ ( self : str ): lowercase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowercase__ = self.get_image_processor(do_normalize=UpperCamelCase__, padding_value=1.0 ) lowercase__ = SamProcessor.from_pretrained(self.tmpdirname, do_normalize=UpperCamelCase__, padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCamelCase__ ) def A__ ( self : Dict ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = self.prepare_image_inputs() lowercase__ = image_processor(UpperCamelCase__, return_tensors="np" ) lowercase__ = processor(images=UpperCamelCase__, return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) @require_torch def A__ ( self : Dict ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = [torch.ones((1, 3, 5, 5) )] lowercase__ = [[1764, 2646]] lowercase__ = [[683, 1024]] lowercase__ = processor.post_process_masks(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) lowercase__ = processor.post_process_masks( UpperCamelCase__, torch.tensor(UpperCamelCase__ ), torch.tensor(UpperCamelCase__ ) ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) # should also work with np lowercase__ = [np.ones((1, 3, 5, 5) )] lowercase__ = processor.post_process_masks(UpperCamelCase__, np.array(UpperCamelCase__ ), np.array(UpperCamelCase__ ) ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) lowercase__ = [[1, 0], [0, 1]] with self.assertRaises(UpperCamelCase__ ): lowercase__ = processor.post_process_masks(UpperCamelCase__, np.array(UpperCamelCase__ ), np.array(UpperCamelCase__ ) ) @require_vision @require_tf class _snake_case ( unittest.TestCase): def A__ ( self : Dict ): lowercase__ = tempfile.mkdtemp() lowercase__ = SamImageProcessor() lowercase__ = SamProcessor(UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def A__ ( self : Union[str, Any], **__lowercase : Optional[int] ): return AutoProcessor.from_pretrained(self.tmpdirname, **UpperCamelCase__ ).image_processor def A__ ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def A__ ( self : List[str] ): lowercase__ = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] lowercase__ = [Image.fromarray(np.moveaxis(UpperCamelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def A__ ( self : List[str] ): lowercase__ = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowercase__ = self.get_image_processor(do_normalize=UpperCamelCase__, padding_value=1.0 ) lowercase__ = SamProcessor.from_pretrained(self.tmpdirname, do_normalize=UpperCamelCase__, padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCamelCase__ ) def A__ ( self : Optional[int] ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = self.prepare_image_inputs() lowercase__ = image_processor(UpperCamelCase__, return_tensors="np" ) lowercase__ = processor(images=UpperCamelCase__, return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) @require_tf def A__ ( self : Any ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = [tf.ones((1, 3, 5, 5) )] lowercase__ = [[1764, 2646]] lowercase__ = [[683, 1024]] lowercase__ = processor.post_process_masks(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, return_tensors="tf" ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) lowercase__ = processor.post_process_masks( UpperCamelCase__, tf.convert_to_tensor(UpperCamelCase__ ), tf.convert_to_tensor(UpperCamelCase__ ), return_tensors="tf", ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) # should also work with np lowercase__ = [np.ones((1, 3, 5, 5) )] lowercase__ = processor.post_process_masks( UpperCamelCase__, np.array(UpperCamelCase__ ), np.array(UpperCamelCase__ ), return_tensors="tf" ) self.assertEqual(masks[0].shape, (1, 3, 1764, 2646) ) lowercase__ = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): lowercase__ = processor.post_process_masks( UpperCamelCase__, np.array(UpperCamelCase__ ), np.array(UpperCamelCase__ ), return_tensors="tf" ) @require_vision @require_torchvision class _snake_case ( unittest.TestCase): def A__ ( self : Any ): lowercase__ = tempfile.mkdtemp() lowercase__ = SamImageProcessor() lowercase__ = SamProcessor(UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) def A__ ( self : Optional[int], **__lowercase : int ): return AutoProcessor.from_pretrained(self.tmpdirname, **UpperCamelCase__ ).image_processor def A__ ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def A__ ( self : Tuple ): lowercase__ = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] lowercase__ = [Image.fromarray(np.moveaxis(UpperCamelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def A__ ( self : Optional[int] ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = np.random.randint(0, 2, size=(1, 3, 5, 5) ).astype(np.floataa ) lowercase__ = [tf.convert_to_tensor(UpperCamelCase__ )] lowercase__ = [torch.tensor(UpperCamelCase__ )] lowercase__ = [[1764, 2646]] lowercase__ = [[683, 1024]] lowercase__ = processor.post_process_masks( UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, return_tensors="tf" ) lowercase__ = processor.post_process_masks( UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def A__ ( self : Optional[int] ): lowercase__ = self.get_image_processor() lowercase__ = SamProcessor(image_processor=UpperCamelCase__ ) lowercase__ = self.prepare_image_inputs() lowercase__ = image_processor(UpperCamelCase__, return_tensors="pt" )['''pixel_values'''].numpy() lowercase__ = processor(images=UpperCamelCase__, return_tensors="pt" )['''pixel_values'''].numpy() lowercase__ = image_processor(UpperCamelCase__, return_tensors="tf" )['''pixel_values'''].numpy() lowercase__ = processor(images=UpperCamelCase__, return_tensors="tf" )['''pixel_values'''].numpy() self.assertTrue(np.allclose(UpperCamelCase__, UpperCamelCase__ ) ) self.assertTrue(np.allclose(UpperCamelCase__, UpperCamelCase__ ) ) self.assertTrue(np.allclose(UpperCamelCase__, UpperCamelCase__ ) )
710
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase__ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowercase__ = f'''{src_lang}-{tgt_lang}''' lowercase__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) lowercase__ = os.path.join(SCREAMING_SNAKE_CASE_ , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
37
0
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ ): lowercase__ = int(number**0.5 ) return number == sq * sq def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den lowercase__ = x_den * y_den * z_den lowercase__ = gcd(lowerCamelCase_ , lowerCamelCase_ ) top //= hcf bottom //= hcf return top, bottom def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ = 35 ): lowercase__ = set() lowercase__ = 42 lowercase__ = Fraction(0 ) lowercase__ = 42 for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 lowercase__ = x_num * y_den + x_den * y_num lowercase__ = x_den * y_den lowercase__ = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowercase__ = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 lowercase__ = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) lowercase__ = x_den * x_den * y_den * y_den if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): lowercase__ = int(sqrt(lowerCamelCase_ ) ) lowercase__ = int(sqrt(lowerCamelCase_ ) ) lowercase__ = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowercase__ = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=-1 lowercase__ = x_num * y_num lowercase__ = x_den * y_num + x_num * y_den lowercase__ = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowercase__ = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 lowercase__ = x_num * x_num * y_num * y_num lowercase__ = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): lowercase__ = int(sqrt(lowerCamelCase_ ) ) lowercase__ = int(sqrt(lowerCamelCase_ ) ) lowercase__ = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowercase__ = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) for num, den in unique_s: total += Fraction(lowerCamelCase_ , lowerCamelCase_ ) return total.denominator + total.numerator if __name__ == "__main__": print(F'{solution() = }')
711
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =TransfoXLTokenizer UpperCamelCase__ : List[Any] =False UpperCamelCase__ : List[Any] =False def A__ ( self : Union[str, Any] ): super().setUp() lowercase__ = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : Union[str, Any], **__lowercase : Any ): lowercase__ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname, **__lowercase ) def A__ ( self : Tuple, __lowercase : Optional[int] ): lowercase__ = "<unk> UNwanted , running" lowercase__ = "<unk> unwanted, running" return input_text, output_text def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(vocab_file=self.vocab_file, lower_case=__lowercase ) lowercase__ = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowercase, ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ), [0, 4, 8, 7] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["hello", "!", "how", "are", "you", "?"] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) lowercase__ = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" lowercase__ = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowercase ), __lowercase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowercase ), __lowercase ) def A__ ( self : List[str] ): lowercase__ = self.get_tokenizer() lowercase__ = len(__lowercase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1", 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowercase ), original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ), [1] ) self.assertEqual(tokenizer.decode([1] ), "new1" )
37
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
712
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser(SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.parse_args_into_dataclasses()[0] lowercase__ = TensorFlowBenchmark(args=SCREAMING_SNAKE_CASE_ ) try: lowercase__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowercase__ = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowercase__ = " ".join(str(SCREAMING_SNAKE_CASE_ ).split(" " )[:-1] ) lowercase__ = "" lowercase__ = eval(str(SCREAMING_SNAKE_CASE_ ).split(" " )[-1] ) lowercase__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = full_error_msg + begin_error_msg + str(SCREAMING_SNAKE_CASE_ ) raise ValueError(SCREAMING_SNAKE_CASE_ ) benchmark.run() if __name__ == "__main__": main()
37
0
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class _snake_case ( __lowercase): UpperCamelCase__ : int =0 UpperCamelCase__ : bool =False UpperCamelCase__ : float =3.0 class _snake_case ( unittest.TestCase): def A__ ( self : Optional[Any] ): # If no defaults are changed, `to_kwargs` returns an empty dict. self.assertDictEqual(MockClass().to_kwargs(), {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs(), {"a": 2} ) self.assertDictEqual(MockClass(a=2, b=__A ).to_kwargs(), {"a": 2, "b": True} ) self.assertDictEqual(MockClass(a=2, c=2.25 ).to_kwargs(), {"a": 2, "c": 2.25} ) @require_cuda def A__ ( self : int ): # If no defaults are changed, `to_kwargs` returns an empty dict. lowercase__ = GradScalerKwargs(init_scale=1024, growth_factor=2 ) AcceleratorState._reset_state() lowercase__ = Accelerator(mixed_precision="fp16", kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) lowercase__ = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale, 1024.0 ) self.assertEqual(scaler._growth_factor, 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor, 0.5 ) self.assertEqual(scaler._growth_interval, 2000 ) self.assertEqual(scaler._enabled, __A ) @require_multi_gpu def A__ ( self : Dict ): lowercase__ = ["torchrun", F'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(__A, env=os.environ.copy() ) if __name__ == "__main__": lowercase_ = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) lowercase_ = Accelerator(kwargs_handlers=[ddp_scaler]) lowercase_ = torch.nn.Linear(100, 200) lowercase_ = accelerator.prepare(model) # Check the values changed in kwargs lowercase_ = """""" lowercase_ = model.bucket_bytes_cap // (1024 * 1024) if observed_bucket_cap_map != 15: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
713
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowercase_ = """<<<<<<< This should probably be modified because it mentions: """ lowercase_ = """======= >>>>>>> """ lowercase_ = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] lowercase_ = [ # (pattern, replacement) # Order is important here for some replacements (r"""tfds\.core""", r"""datasets"""), (r"""tf\.io\.gfile\.GFile""", r"""open"""), (r"""tf\.([\w\d]+)""", r"""datasets.Value('\1')"""), (r"""tfds\.features\.Text\(\)""", r"""datasets.Value('string')"""), (r"""tfds\.features\.Text\(""", r"""datasets.Value('string'),"""), (r"""features\s*=\s*tfds.features.FeaturesDict\(""", r"""features=datasets.Features("""), (r"""tfds\.features\.FeaturesDict\(""", r"""dict("""), (r"""The TensorFlow Datasets Authors""", r"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (r"""tfds\.""", r"""datasets."""), (r"""dl_manager\.manual_dir""", r"""self.config.data_dir"""), (r"""self\.builder_config""", r"""self.config"""), ] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return ConvertCommand(args.tfds_path , args.datasets_directory ) class _snake_case ( lowercase__): @staticmethod def A__ ( __lowercase : ArgumentParser ): lowercase__ = parser.add_parser( "convert", help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.", ) train_parser.add_argument( "--tfds_path", type=__lowercase, required=__lowercase, help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.", ) train_parser.add_argument( "--datasets_directory", type=__lowercase, required=__lowercase, help="Path to the HuggingFace Datasets folder." ) train_parser.set_defaults(func=__lowercase ) def __init__( self : Tuple, __lowercase : str, __lowercase : str, *__lowercase : Tuple ): lowercase__ = get_logger("datasets-cli/converting" ) lowercase__ = tfds_path lowercase__ = datasets_directory def A__ ( self : Any ): if os.path.isdir(self._tfds_path ): lowercase__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__ = os.path.dirname(self._tfds_path ) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path." ) lowercase__ = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) lowercase__ = [] lowercase__ = [] lowercase__ = {} if os.path.isdir(self._tfds_path ): lowercase__ = os.listdir(__lowercase ) else: lowercase__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) if not os.path.isfile(__lowercase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file" ) continue with open(__lowercase, encoding="utf-8" ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = False lowercase__ = False lowercase__ = [] for line in lines: lowercase__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__ = "import datasets\n" elif "import tensorflow" in out_line: # order is important here lowercase__ = "" continue elif "from absl import logging" in out_line: lowercase__ = "from datasets import logging\n" elif "getLogger" in out_line: lowercase__ = out_line.replace("getLogger", "get_logger" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__ = True lowercase__ = list(filter(lambda __lowercase : e in out_line, __lowercase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowercase ) + "\n" ) out_lines.append(__lowercase ) out_lines.append(__lowercase ) continue else: for pattern, replacement in TO_CONVERT: lowercase__ = re.sub(__lowercase, __lowercase, __lowercase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__ = re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)", __lowercase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split("," ) ) lowercase__ = "from . import " + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__ = True out_lines.append(__lowercase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__ = f_name.replace(".py", "" ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) os.makedirs(__lowercase, exist_ok=__lowercase ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowercase ) if needs_manual_update: with_manual_update.append(__lowercase ) with open(__lowercase, "w", encoding="utf-8" ) as f: f.writelines(__lowercase ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: lowercase__ = os.path.basename(__lowercase ) lowercase__ = imports_to_builder_map[f_name.replace(".py", "" )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowercase, __lowercase ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
37
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = 5000_0000 ): lowercase__ = set() lowercase__ = int((limit - 24) ** (1 / 2) ) lowercase__ = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , UpperCamelCase__ ) ) ) for primea in primes: lowercase__ = primea * primea for primea in primes: lowercase__ = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: lowercase__ = primea * primea * primea * primea lowercase__ = square + cube + tetr if total >= limit: break ret.add(UpperCamelCase__ ) return len(UpperCamelCase__ ) if __name__ == "__main__": print(F'{solution() = }')
714
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase_ = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } lowercase_ = { """allenai/led-base-16384""": 1_6384, } class _snake_case ( lowercase__): UpperCamelCase__ : int =VOCAB_FILES_NAMES UpperCamelCase__ : Any =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =LEDTokenizer UpperCamelCase__ : Tuple =["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any], __lowercase : Optional[Any]=None, __lowercase : Dict=None, __lowercase : Tuple=None, __lowercase : Union[str, Any]="replace", __lowercase : Tuple="<s>", __lowercase : Optional[Any]="</s>", __lowercase : Tuple="</s>", __lowercase : List[str]="<s>", __lowercase : Tuple="<unk>", __lowercase : Dict="<pad>", __lowercase : Dict="<mask>", __lowercase : Any=False, __lowercase : Any=True, **__lowercase : List[Any], ): super().__init__( __lowercase, __lowercase, tokenizer_file=__lowercase, errors=__lowercase, bos_token=__lowercase, eos_token=__lowercase, sep_token=__lowercase, cls_token=__lowercase, unk_token=__lowercase, pad_token=__lowercase, mask_token=__lowercase, add_prefix_space=__lowercase, trim_offsets=__lowercase, **__lowercase, ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = getattr(__lowercase, pre_tok_state.pop("type" ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**__lowercase ) lowercase__ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase__ = "post_processor" lowercase__ = getattr(self.backend_tokenizer, __lowercase, __lowercase ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state["sep"] ) if "cls" in state: lowercase__ = tuple(state["cls"] ) lowercase__ = False if state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get("trim_offsets", __lowercase ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(__lowercase, state.pop("type" ) ) lowercase__ = component_class(**__lowercase ) setattr(self.backend_tokenizer, __lowercase, __lowercase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A__ ( self : str ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self : Optional[int], __lowercase : Dict ): lowercase__ = AddedToken(__lowercase, lstrip=__lowercase, rstrip=__lowercase ) if isinstance(__lowercase, __lowercase ) else value lowercase__ = value def A__ ( self : Any, *__lowercase : List[Any], **__lowercase : Optional[Any] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowercase, **__lowercase ) def A__ ( self : int, *__lowercase : Union[str, Any], **__lowercase : List[str] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowercase, **__lowercase ) def A__ ( self : Optional[Any], __lowercase : str, __lowercase : Optional[str] = None ): lowercase__ = self._tokenizer.model.save(__lowercase, name=__lowercase ) return tuple(__lowercase ) def A__ ( self : List[str], __lowercase : int, __lowercase : Optional[int]=None ): lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self : int, __lowercase : List[int], __lowercase : Optional[List[int]] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self : Union[str, Any], __lowercase : Union[Dict[str, EncodedInput], BatchEncoding], __lowercase : Optional[int] = None, __lowercase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD, __lowercase : Optional[int] = None, __lowercase : Optional[bool] = None, ): lowercase__ = super()._pad( encoded_inputs=__lowercase, max_length=__lowercase, padding_strategy=__lowercase, pad_to_multiple_of=__lowercase, return_attention_mask=__lowercase, ) # Load from model defaults if return_attention_mask is None: lowercase__ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ = len(encoded_inputs["global_attention_mask"] ) != len(__lowercase ) if needs_to_be_padded: lowercase__ = len(__lowercase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
37
0
from manim import * class _snake_case ( __UpperCAmelCase): def A__ ( self : List[Any] ): lowercase__ = Rectangle(height=0.5, width=0.5 ) lowercase__ = Rectangle(height=0.25, width=0.25 ) lowercase__ = Rectangle(height=0.46, width=0.46 ).set_stroke(width=0 ) lowercase__ = [mem.copy() for i in range(6 )] lowercase__ = [mem.copy() for i in range(6 )] lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = VGroup(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = Text("CPU", font_size=24 ) lowercase__ = Group(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0.5, aligned_edge=__SCREAMING_SNAKE_CASE ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ = [mem.copy() for i in range(4 )] lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = Text("GPU", font_size=24 ) lowercase__ = Group(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0.5, aligned_edge=__SCREAMING_SNAKE_CASE ) gpu.move_to([-1, -1, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ = [mem.copy() for i in range(6 )] lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = Text("Model", font_size=24 ) lowercase__ = Group(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0.5, aligned_edge=__SCREAMING_SNAKE_CASE ) model.move_to([3, -1.0, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ = [] lowercase__ = [] lowercase__ = [] for i, rect in enumerate(__SCREAMING_SNAKE_CASE ): rect.set_stroke(__SCREAMING_SNAKE_CASE ) lowercase__ = Rectangle(height=0.46 / 4, width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__SCREAMING_SNAKE_CASE, opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ), buff=0.02, direction=__SCREAMING_SNAKE_CASE ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0], direction=__SCREAMING_SNAKE_CASE, buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1], direction=__SCREAMING_SNAKE_CASE, buff=0.0 ) self.add(__SCREAMING_SNAKE_CASE ) model_cpu_arr.append(__SCREAMING_SNAKE_CASE ) self.add(*__SCREAMING_SNAKE_CASE, *__SCREAMING_SNAKE_CASE, *__SCREAMING_SNAKE_CASE ) lowercase__ = [mem.copy() for i in range(6 )] lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = Text("Loaded Checkpoint", font_size=24 ) lowercase__ = Group(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0.5, aligned_edge=__SCREAMING_SNAKE_CASE ) checkpoint.move_to([3, 0.5, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ = [] lowercase__ = [] for i, rect in enumerate(__SCREAMING_SNAKE_CASE ): lowercase__ = fill.copy().set_fill(__SCREAMING_SNAKE_CASE, opacity=0.7 ) target.move_to(__SCREAMING_SNAKE_CASE ) ckpt_arr.append(__SCREAMING_SNAKE_CASE ) lowercase__ = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(__SCREAMING_SNAKE_CASE ) self.add(*__SCREAMING_SNAKE_CASE, *__SCREAMING_SNAKE_CASE ) lowercase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowercase__ = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''', font_size=18, ) key_text.move_to([-5, 2.4, 0] ) self.add(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ) lowercase__ = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''', font_size=18, ) blue_text.next_to(__SCREAMING_SNAKE_CASE, DOWN * 2.4, aligned_edge=key_text.get_left() ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ = MarkupText( F'''Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.''', font_size=24, ) step_a.move_to([2, 2, 0] ) lowercase__ = [meta_mem.copy() for i in range(6 )] lowercase__ = [meta_mem.copy() for i in range(6 )] lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = VGroup(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0 ) lowercase__ = Text("Disk", font_size=24 ) lowercase__ = Group(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE, buff=0.5, aligned_edge=__SCREAMING_SNAKE_CASE ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(__SCREAMING_SNAKE_CASE, run_time=3 ), Write(__SCREAMING_SNAKE_CASE, run_time=1 ), Create(__SCREAMING_SNAKE_CASE, run_time=1 ) ) lowercase__ = [] for i, rect in enumerate(__SCREAMING_SNAKE_CASE ): lowercase__ = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(__SCREAMING_SNAKE_CASE, run_time=1.5 ) ) self.play(*__SCREAMING_SNAKE_CASE ) self.play(FadeOut(__SCREAMING_SNAKE_CASE ) ) lowercase__ = MarkupText(F'''Then, the checkpoint is removed from memory\nthrough garbage collection.''', font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(__SCREAMING_SNAKE_CASE, run_time=3 ) ) self.play( FadeOut(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, *__SCREAMING_SNAKE_CASE, *__SCREAMING_SNAKE_CASE ), ) self.wait()
715
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __lowerCAmelCase ( ): lowercase__ = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=SCREAMING_SNAKE_CASE_ ) env_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) launch_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) tpu_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) test_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) # Let's go lowercase__ = parser.parse_args() if not hasattr(SCREAMING_SNAKE_CASE_ , "func" ): parser.print_help() exit(1 ) # Run args.func(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
37
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = False while is_sorted is False: # Until all the indices are traversed keep looping lowercase__ = True for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) - 1 , 2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: lowercase__ = input_list[i + 1], input_list[i] # swapping if elements not in order lowercase__ = False for i in range(1 , len(SCREAMING_SNAKE_CASE_ ) - 1 , 2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: lowercase__ = input_list[i + 1], input_list[i] # swapping if elements not in order lowercase__ = False return input_list if __name__ == "__main__": print("""Enter list to be sorted""") lowercase_ = [int(x) for x in input().split()] # inputing elements of the list in one line lowercase_ = odd_even_sort(input_list) print("""The sorted list is""") print(sorted_list)
716
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : Dict, __lowercase : int, __lowercase : Union[str, Any]=7, __lowercase : Union[str, Any]=3, __lowercase : Any=18, __lowercase : Union[str, Any]=30, __lowercase : Any=400, __lowercase : List[str]=True, __lowercase : Dict=None, __lowercase : List[str]=True, __lowercase : int=False, __lowercase : Union[str, Any]=True, __lowercase : str=True, __lowercase : Optional[int]=[0.5, 0.5, 0.5], __lowercase : List[Any]=[0.5, 0.5, 0.5], ): lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size if size is not None else {"height": 18, "width": 20} lowercase__ = do_thumbnail lowercase__ = do_align_axis lowercase__ = do_pad lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std def A__ ( self : Optional[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =DonutImageProcessor if is_vision_available() else None def A__ ( self : str ): lowercase__ = DonutImageProcessingTester(self ) @property def A__ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_resize" ) ) self.assertTrue(hasattr(__lowercase, "size" ) ) self.assertTrue(hasattr(__lowercase, "do_thumbnail" ) ) self.assertTrue(hasattr(__lowercase, "do_align_long_axis" ) ) self.assertTrue(hasattr(__lowercase, "do_pad" ) ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "image_mean" ) ) self.assertTrue(hasattr(__lowercase, "image_std" ) ) def A__ ( self : str ): lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {"height": 18, "width": 20} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {"height": 42, "width": 42} ) # Previous config had dimensions in (width, height) order lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=(42, 84) ) self.assertEqual(image_processor.size, {"height": 84, "width": 42} ) def A__ ( self : List[str] ): pass @is_flaky() def A__ ( self : Dict ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Optional[Any] ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Tuple ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), )
37
0
from ...processing_utils import ProcessorMixin class _snake_case ( lowercase__): UpperCamelCase__ : Optional[int] =["""image_processor""", """feature_extractor"""] UpperCamelCase__ : Union[str, Any] ="""TvltImageProcessor""" UpperCamelCase__ : List[Any] ="""TvltFeatureExtractor""" def __init__( self : List[str], __lowercase : List[Any], __lowercase : List[str] ): super().__init__(image_processor=__A, feature_extractor=__A ) lowercase__ = image_processor lowercase__ = feature_extractor def __call__( self : List[str], __lowercase : Any=None, __lowercase : Any=None, __lowercase : Optional[int]=None, __lowercase : str=None, __lowercase : List[Any]=False, __lowercase : Tuple=False, *__lowercase : List[str], **__lowercase : Optional[Any], ): if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) lowercase__ = None if images is not None: lowercase__ = self.image_processor(__A, mask_pixel=__A, *__A, **__A ) if images_mixed is not None: lowercase__ = self.image_processor(__A, is_mixed=__A, *__A, **__A ) if audio is not None: lowercase__ = self.feature_extractor( __A, *__A, sampling_rate=__A, mask_audio=__A, **__A ) lowercase__ = {} if audio is not None: output_dict.update(__A ) if images is not None: output_dict.update(__A ) if images_mixed_dict is not None: output_dict.update(__A ) return output_dict @property def A__ ( self : Optional[Any] ): lowercase__ = self.image_processor.model_input_names lowercase__ = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
717
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _snake_case ( lowercase__): def A__ ( self : Optional[Any], __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) lowercase__ = input_file.read() lowercase__ = regexp.search(__lowercase ) return match def A__ ( self : str, __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()", re.DOTALL ) lowercase__ = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowercase__ = regexp.finditer(__lowercase ) lowercase__ = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(F'''open(...) must use utf-8 encoding in {dataset}''' ) def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(F'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
37
0
import functools def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): # Validation if not isinstance(_lowerCamelCase , _lowerCamelCase ) or not all(isinstance(_lowerCamelCase , _lowerCamelCase ) for day in days ): raise ValueError("The parameter days should be a list of integers" ) if len(_lowerCamelCase ) != 3 or not all(isinstance(_lowerCamelCase , _lowerCamelCase ) for cost in costs ): raise ValueError("The parameter costs should be a list of three integers" ) if len(_lowerCamelCase ) == 0: return 0 if min(_lowerCamelCase ) <= 0: raise ValueError("All days elements should be greater than 0" ) if max(_lowerCamelCase ) >= 366: raise ValueError("All days elements should be less than 366" ) lowercase__ = set(_lowerCamelCase ) @functools.cache def dynamic_programming(SCREAMING_SNAKE_CASE_ ) -> int: if index > 365: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 30 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
718
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
lowercase_ = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on lowercase_ = {value: key for key, value in MORSE_CODE_DICT.items()} def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return "".join(REVERSE_DICT[char] for char in message.split() ) def __lowerCAmelCase ( ): lowercase__ = """Morse code here!""" print(lowercase_ ) lowercase__ = encrypt(lowercase_ ) print(lowercase_ ) lowercase__ = decrypt(lowercase_ ) print(lowercase_ ) if __name__ == "__main__": main()
719
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowercase_ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : List[Any], __lowercase : int, __lowercase : Optional[int]=7, __lowercase : List[str]=3, __lowercase : Tuple=18, __lowercase : List[Any]=30, __lowercase : Tuple=400, __lowercase : Any=None, __lowercase : Optional[int]=True, __lowercase : List[str]=True, __lowercase : Union[str, Any]=None, ): lowercase__ = size if size is not None else {"height": 20, "width": 20} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = size lowercase__ = do_normalize lowercase__ = do_convert_rgb lowercase__ = [512, 1024, 2048, 4096] lowercase__ = patch_size if patch_size is not None else {"height": 16, "width": 16} def A__ ( self : List[str] ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def A__ ( self : Any ): lowercase__ = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" lowercase__ = Image.open(requests.get(__lowercase, stream=__lowercase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Any =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Any ): lowercase__ = PixaStructImageProcessingTester(self ) @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Optional[int] ): lowercase__ = self.image_processor_tester.prepare_dummy_image() lowercase__ = self.image_processing_class(**self.image_processor_dict ) lowercase__ = 2048 lowercase__ = image_processor(__lowercase, return_tensors="pt", max_patches=__lowercase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean(), torch.tensor(0.0606 ), atol=1e-3, rtol=1e-3 ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : int ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 lowercase__ = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__lowercase ): lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches lowercase__ = "Hello" lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Tuple ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Any ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Optional[int] ): lowercase__ = PixaStructImageProcessingTester(self, num_channels=4 ) lowercase__ = 3 @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), )
37
0
'''simple docstring''' import random def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ , lowercase__ = [], [], [] for element in data: if element < pivot: less.append(lowerCAmelCase__ ) elif element > pivot: greater.append(lowerCAmelCase__ ) else: equal.append(lowerCAmelCase__ ) return less, equal, greater def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(lowerCAmelCase__ ) or index < 0: return None lowercase__ = items[random.randint(0 , len(lowerCAmelCase__ ) - 1 )] lowercase__ = 0 lowercase__ , lowercase__ , lowercase__ = _partition(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__ = len(lowerCAmelCase__ ) lowercase__ = len(lowerCAmelCase__ ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(lowerCAmelCase__ , lowerCAmelCase__ ) # must be in larger else: return quick_select(lowerCAmelCase__ , index - (m + count) )
720
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = len(SCREAMING_SNAKE_CASE_ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ = 0 count += depth_first_search(SCREAMING_SNAKE_CASE_ , row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , row - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col - 1 , SCREAMING_SNAKE_CASE_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
37
0
from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class _snake_case : def __init__( self : List[str], __lowercase : List[Any], ): lowercase__ = parent lowercase__ = 13 lowercase__ = 7 lowercase__ = True lowercase__ = True lowercase__ = False lowercase__ = True lowercase__ = 99 lowercase__ = 32 lowercase__ = 2 lowercase__ = 4 lowercase__ = 37 lowercase__ = "gelu" lowercase__ = 0.1 lowercase__ = 0.1 lowercase__ = 512 lowercase__ = 16 lowercase__ = 2 lowercase__ = 0.02 lowercase__ = 3 lowercase__ = 4 lowercase__ = None def A__ ( self : str ): lowercase__ = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) lowercase__ = None if self.use_input_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ = None lowercase__ = None lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size], self.type_sequence_label_size ) lowercase__ = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) lowercase__ = ids_tensor([self.batch_size], self.num_choices ) lowercase__ = DistilBertConfig( vocab_size=self.vocab_size, dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, hidden_dim=self.intermediate_size, hidden_act=self.hidden_act, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self : Optional[Any], __lowercase : str, __lowercase : int, __lowercase : str, __lowercase : List[Any], __lowercase : Optional[int], __lowercase : str ): lowercase__ = TFDistilBertModel(config=UpperCamelCase__ ) lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowercase__ = model(UpperCamelCase__ ) lowercase__ = [input_ids, input_mask] lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self : int, __lowercase : Tuple, __lowercase : Tuple, __lowercase : Union[str, Any], __lowercase : Any, __lowercase : Optional[int], __lowercase : int ): lowercase__ = TFDistilBertForMaskedLM(config=UpperCamelCase__ ) lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def A__ ( self : Optional[Any], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : Dict, __lowercase : List[str], __lowercase : Dict, __lowercase : Any ): lowercase__ = TFDistilBertForQuestionAnswering(config=UpperCamelCase__ ) lowercase__ = { "input_ids": input_ids, "attention_mask": input_mask, } lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def A__ ( self : str, __lowercase : Union[str, Any], __lowercase : Optional[int], __lowercase : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Tuple ): lowercase__ = self.num_labels lowercase__ = TFDistilBertForSequenceClassification(UpperCamelCase__ ) lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def A__ ( self : List[Any], __lowercase : Optional[Any], __lowercase : Tuple, __lowercase : Dict, __lowercase : int, __lowercase : List[str], __lowercase : Any ): lowercase__ = self.num_choices lowercase__ = TFDistilBertForMultipleChoice(UpperCamelCase__ ) lowercase__ = tf.tile(tf.expand_dims(UpperCamelCase__, 1 ), (1, self.num_choices, 1) ) lowercase__ = tf.tile(tf.expand_dims(UpperCamelCase__, 1 ), (1, self.num_choices, 1) ) lowercase__ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices) ) def A__ ( self : str, __lowercase : List[str], __lowercase : Tuple, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : int ): lowercase__ = self.num_labels lowercase__ = TFDistilBertForTokenClassification(UpperCamelCase__ ) lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} lowercase__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def A__ ( self : Dict ): lowercase__ = self.prepare_config_and_inputs() ((lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__)) = config_and_inputs lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _snake_case ( lowercase__ , lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) UpperCamelCase__ : Union[str, Any] =( { "feature-extraction": TFDistilBertModel, "fill-mask": TFDistilBertForMaskedLM, "question-answering": TFDistilBertForQuestionAnswering, "text-classification": TFDistilBertForSequenceClassification, "token-classification": TFDistilBertForTokenClassification, "zero-shot": TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) UpperCamelCase__ : List[Any] =False UpperCamelCase__ : int =False def A__ ( self : str ): lowercase__ = TFDistilBertModelTester(self ) lowercase__ = ConfigTester(self, config_class=UpperCamelCase__, dim=37 ) def A__ ( self : Optional[Any] ): self.config_tester.run_common_tests() def A__ ( self : str ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*UpperCamelCase__ ) def A__ ( self : List[str] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*UpperCamelCase__ ) def A__ ( self : int ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*UpperCamelCase__ ) def A__ ( self : Any ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*UpperCamelCase__ ) def A__ ( self : int ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*UpperCamelCase__ ) def A__ ( self : List[str] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*UpperCamelCase__ ) @slow def A__ ( self : List[Any] ): for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): lowercase__ = TFDistilBertModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) @require_tf class _snake_case ( unittest.TestCase): @slow def A__ ( self : Union[str, Any] ): lowercase__ = TFDistilBertModel.from_pretrained("distilbert-base-uncased" ) lowercase__ = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase__ = model(UpperCamelCase__ )[0] lowercase__ = [1, 6, 768] self.assertEqual(output.shape, UpperCamelCase__ ) lowercase__ = tf.constant( [ [ [0.19261885, -0.13732955, 0.4119799], [0.22150156, -0.07422661, 0.39037204], [0.22756018, -0.0896414, 0.3701467], ] ] ) tf.debugging.assert_near(output[:, :3, :3], UpperCamelCase__, atol=1e-4 )
721
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 for ch in input_str: lowercase__ = ord(SCREAMING_SNAKE_CASE_ ) lowercase__ = pow(2 , SCREAMING_SNAKE_CASE_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
37
0
'''simple docstring''' import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """kakaobrain/align-base""": """https://huggingface.co/kakaobrain/align-base/resolve/main/config.json""", } class _snake_case ( snake_case_): UpperCamelCase__ : Optional[Any] ="""align_text_model""" def __init__( self : Dict, __lowercase : Optional[Any]=3_0522, __lowercase : str=768, __lowercase : Any=12, __lowercase : Optional[Any]=12, __lowercase : Dict=3072, __lowercase : List[Any]="gelu", __lowercase : Union[str, Any]=0.1, __lowercase : Dict=0.1, __lowercase : List[Any]=512, __lowercase : str=2, __lowercase : Dict=0.02, __lowercase : str=1e-1_2, __lowercase : List[Any]=0, __lowercase : List[Any]="absolute", __lowercase : int=True, **__lowercase : Optional[Any], ): super().__init__(**__lowercase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = pad_token_id @classmethod def A__ ( cls : int, __lowercase : Optional[Any], **__lowercase : Dict ): cls._set_token_in_kwargs(__lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(__lowercase, **__lowercase ) # get the text config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": lowercase__ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls, "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__lowercase, **__lowercase ) class _snake_case ( snake_case_): UpperCamelCase__ : Dict ="""align_vision_model""" def __init__( self : Optional[int], __lowercase : List[str] = 3, __lowercase : str = 600, __lowercase : Optional[int] = 2.0, __lowercase : Any = 3.1, __lowercase : Any = 8, __lowercase : Any = [3, 3, 5, 3, 5, 5, 3], __lowercase : Tuple = [32, 16, 24, 40, 80, 112, 192], __lowercase : List[str] = [16, 24, 40, 80, 112, 192, 320], __lowercase : int = [], __lowercase : List[Any] = [1, 2, 2, 2, 1, 2, 1], __lowercase : int = [1, 2, 2, 3, 3, 4, 1], __lowercase : Any = [1, 6, 6, 6, 6, 6, 6], __lowercase : Optional[Any] = 0.25, __lowercase : int = "swish", __lowercase : List[Any] = 2560, __lowercase : Any = "mean", __lowercase : int = 0.02, __lowercase : Any = 0.001, __lowercase : int = 0.99, __lowercase : str = 0.2, **__lowercase : List[Any], ): super().__init__(**__lowercase ) lowercase__ = num_channels lowercase__ = image_size lowercase__ = width_coefficient lowercase__ = depth_coefficient lowercase__ = depth_divisor lowercase__ = kernel_sizes lowercase__ = in_channels lowercase__ = out_channels lowercase__ = depthwise_padding lowercase__ = strides lowercase__ = num_block_repeats lowercase__ = expand_ratios lowercase__ = squeeze_expansion_ratio lowercase__ = hidden_act lowercase__ = hidden_dim lowercase__ = pooling_type lowercase__ = initializer_range lowercase__ = batch_norm_eps lowercase__ = batch_norm_momentum lowercase__ = drop_connect_rate lowercase__ = sum(__lowercase ) * 4 @classmethod def A__ ( cls : str, __lowercase : str, **__lowercase : Optional[Any] ): cls._set_token_in_kwargs(__lowercase ) lowercase__ , lowercase__ = cls.get_config_dict(__lowercase, **__lowercase ) # get the vision config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": lowercase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls, "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__lowercase, **__lowercase ) class _snake_case ( snake_case_): UpperCamelCase__ : str ="""align""" UpperCamelCase__ : List[Any] =True def __init__( self : Optional[int], __lowercase : Optional[Any]=None, __lowercase : List[str]=None, __lowercase : int=640, __lowercase : Union[str, Any]=1.0, __lowercase : List[Any]=0.02, **__lowercase : List[Any], ): super().__init__(**__lowercase ) if text_config is None: lowercase__ = {} logger.info("text_config is None. Initializing the AlignTextConfig with default values." ) if vision_config is None: lowercase__ = {} logger.info("vision_config is None. Initializing the AlignVisionConfig with default values." ) lowercase__ = AlignTextConfig(**__lowercase ) lowercase__ = AlignVisionConfig(**__lowercase ) lowercase__ = projection_dim lowercase__ = temperature_init_value lowercase__ = initializer_range @classmethod def A__ ( cls : Dict, __lowercase : List[str], __lowercase : Union[str, Any], **__lowercase : int ): return cls(text_config=text_config.to_dict(), vision_config=vision_config.to_dict(), **__lowercase ) def A__ ( self : Optional[Any] ): lowercase__ = copy.deepcopy(self.__dict__ ) lowercase__ = self.text_config.to_dict() lowercase__ = self.vision_config.to_dict() lowercase__ = self.__class__.model_type return output
700
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = len(SCREAMING_SNAKE_CASE_ ) for i in range(SCREAMING_SNAKE_CASE_ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE_ ): if numbers[j] < numbers[i]: lowercase__ , lowercase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() lowercase_ = [int(item) for item in user_input.split(""",""")] print(exchange_sort(unsorted))
37
0
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __lowerCAmelCase ( *SCREAMING_SNAKE_CASE_ ): if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = list(SCREAMING_SNAKE_CASE_ ) for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowercase__ = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = 128 ): if function is None: return functools.partial(SCREAMING_SNAKE_CASE_ , starting_batch_size=SCREAMING_SNAKE_CASE_ ) lowercase__ = starting_batch_size def decorator(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() lowercase__ = list(inspect.signature(SCREAMING_SNAKE_CASE_ ).parameters.keys() ) # Guard against user error if len(SCREAMING_SNAKE_CASE_ ) < (len(SCREAMING_SNAKE_CASE_ ) + 1): lowercase__ = ", ".join([f'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( f'''Batch size was passed into `{function.__name__}` as the first argument when called.''' f'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(SCREAMING_SNAKE_CASE_ , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) except Exception as e: if should_reduce_batch_size(SCREAMING_SNAKE_CASE_ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
701
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) lowercase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase__ = 1 if upper_limit > 0: lowercase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: lowercase_ = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F'The Catalan numbers from 0 through {N} are:') print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
37
0
from __future__ import annotations from random import random from typing import Generic, TypeVar lowercase_ = TypeVar("""KT""") lowercase_ = TypeVar("""VT""") class _snake_case ( Generic[KT, VT]): def __init__( self : List[Any], __lowercase : KT | str = "root", __lowercase : VT | None = None ): lowercase__ = key lowercase__ = value lowercase__ = [] def __repr__( self : Dict ): return F'''Node({self.key}: {self.value})''' @property def A__ ( self : Optional[Any] ): return len(self.forward ) class _snake_case ( Generic[KT, VT]): def __init__( self : List[str], __lowercase : float = 0.5, __lowercase : int = 16 ): lowercase__ = Node[KT, VT]() lowercase__ = 0 lowercase__ = p lowercase__ = max_level def __str__( self : str ): lowercase__ = list(self ) if len(__lowercase ) == 0: return F'''SkipList(level={self.level})''' lowercase__ = max((len(str(__lowercase ) ) for item in items), default=4 ) lowercase__ = max(__lowercase, 4 ) + 4 lowercase__ = self.head lowercase__ = [] lowercase__ = node.forward.copy() lines.append(F'''[{node.key}]'''.ljust(__lowercase, "-" ) + "* " * len(__lowercase ) ) lines.append(" " * label_size + "| " * len(__lowercase ) ) while len(node.forward ) != 0: lowercase__ = node.forward[0] lines.append( F'''[{node.key}]'''.ljust(__lowercase, "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(__lowercase ) ) lowercase__ = node.forward lines.append("None".ljust(__lowercase ) + "* " * len(__lowercase ) ) return F'''SkipList(level={self.level})\n''' + "\n".join(__lowercase ) def __iter__( self : Optional[int] ): lowercase__ = self.head while len(node.forward ) != 0: yield node.forward[0].key lowercase__ = node.forward[0] def A__ ( self : int ): lowercase__ = 1 while random() < self.p and level < self.max_level: level += 1 return level def A__ ( self : Dict, __lowercase : List[str] ): lowercase__ = [] lowercase__ = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: lowercase__ = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(__lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def A__ ( self : List[str], __lowercase : KT ): lowercase__ = self._locate_node(__lowercase ) if node is not None: for i, update_node in enumerate(__lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: lowercase__ = node.forward[i] else: lowercase__ = update_node.forward[:i] def A__ ( self : List[str], __lowercase : KT, __lowercase : VT ): lowercase__ = self._locate_node(__lowercase ) if node is not None: lowercase__ = value else: lowercase__ = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1, __lowercase ): update_vector.append(self.head ) lowercase__ = level lowercase__ = Node(__lowercase, __lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(__lowercase ) else: lowercase__ = new_node def A__ ( self : List[str], __lowercase : VT ): lowercase__ = self._locate_node(__lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) lowercase__ = skip_list.head lowercase__ = {} while node.level != 0: lowercase__ = node.forward[0] lowercase__ = node.value assert len(SCREAMING_SNAKE_CASE_ ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) lowercase__ = skip_list.head lowercase__ = {} while node.level != 0: lowercase__ = node.forward[0] lowercase__ = node.value if len(SCREAMING_SNAKE_CASE_ ) != 4: print() assert len(SCREAMING_SNAKE_CASE_ ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def __lowerCAmelCase ( ): lowercase__ = SkipList() assert skip_list.find("Some key" ) is None def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(SCREAMING_SNAKE_CASE_ ): yield node.key for forward_node in node.forward: yield from traverse_keys(SCREAMING_SNAKE_CASE_ ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): def is_sorted(SCREAMING_SNAKE_CASE_ ): return all(next_item >= item for item, next_item in zip(SCREAMING_SNAKE_CASE_ , lst[1:] ) ) lowercase__ = SkipList() for i in range(10 ): skip_list.insert(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) def __lowerCAmelCase ( ): for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): lowercase__ = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
702
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase_ = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class _snake_case ( __A): def __lt__( self : Any, __lowercase : List[str] ): return self[-1] < other[-1] def __eq__( self : List[Any], __lowercase : int ): return self[-1] == other[-1] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = [] # sort into stacks for element in collection: lowercase__ = Stack([element] ) lowercase__ = bisect_left(__lowercase , __lowercase ) if i != len(__lowercase ): stacks[i].append(__lowercase ) else: stacks.append(__lowercase ) # use a heap-based merge to merge stack efficiently lowercase__ = merge(*(reversed(__lowercase ) for stack in stacks) ) return collection if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() lowercase_ = [int(item) for item in user_input.split(""",""")] print(patience_sort(unsorted))
703
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") lowercase__ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f'''bert/{name}''' def create_tf_var(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = session.run(SCREAMING_SNAKE_CASE_ ) print(f'''Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace("-" , "_" ) + ".ckpt" ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_=None ): lowercase__ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory in which to save tensorflow model" ) lowercase__ = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
37
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu lowercase_ = get_tests_dir() + """/test_data/fsmt/fsmt_val_data.json""" with io.open(filename, """r""", encoding="""utf-8""") as f: lowercase_ = json.load(f) @require_torch class _snake_case ( unittest.TestCase): def A__ ( self : Optional[int], __lowercase : List[str] ): return FSMTTokenizer.from_pretrained(_A ) def A__ ( self : str, __lowercase : str ): lowercase__ = FSMTForConditionalGeneration.from_pretrained(_A ).to(_A ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ["en-ru", 26.0], ["ru-en", 22.0], ["en-de", 22.0], ["de-en", 29.0], ] ) @slow def A__ ( self : Union[str, Any], __lowercase : Any, __lowercase : str ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality lowercase__ = F'''facebook/wmt19-{pair}''' lowercase__ = self.get_tokenizer(_A ) lowercase__ = self.get_model(_A ) lowercase__ = bleu_data[pair]["src"] lowercase__ = bleu_data[pair]["tgt"] lowercase__ = tokenizer(_A, return_tensors="pt", truncation=_A, padding="longest" ).to(_A ) lowercase__ = model.generate( input_ids=batch.input_ids, num_beams=8, ) lowercase__ = tokenizer.batch_decode( _A, skip_special_tokens=_A, clean_up_tokenization_spaces=_A ) lowercase__ = calculate_bleu(_A, _A ) print(_A ) self.assertGreaterEqual(scores["bleu"], _A )
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_timesformer""": ["""TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimesformerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimesformerModel""", """TimesformerForVideoClassification""", """TimesformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowercase_ = '''src/diffusers''' lowercase_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. lowercase_ = importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) lowercase_ = spec.loader.load_module() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return line.startswith(lowerCAmelCase__ ) or len(lowerCAmelCase__ ) <= 1 or re.search(r"^\s*\)(\s*->.*:|:)\s*$" , lowerCAmelCase__ ) is not None def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = object_name.split("." ) lowercase__ = 0 # First let's find the module where our object lives. lowercase__ = parts[i] while i < len(lowerCAmelCase__ ) and not os.path.isfile(os.path.join(lowerCAmelCase__ , f'''{module}.py''' ) ): i += 1 if i < len(lowerCAmelCase__ ): lowercase__ = os.path.join(lowerCAmelCase__ , parts[i] ) if i >= len(lowerCAmelCase__ ): raise ValueError(f'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(lowerCAmelCase__ , f'''{module}.py''' ) , "r" , encoding="utf-8" , newline="\n" ) as f: lowercase__ = f.readlines() # Now let's find the class / func in the code! lowercase__ = "" lowercase__ = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCAmelCase__ ) and re.search(rf'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCAmelCase__ ): raise ValueError(f''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowercase__ = line_index while line_index < len(lowerCAmelCase__ ) and _should_continue(lines[line_index] , lowerCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase__ = lines[start_index:line_index] return "".join(lowerCAmelCase__ ) lowercase_ = re.compile(r"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") lowercase_ = re.compile(r"""^\s*(\S+)->(\S+)(\s+.*|$)""") lowercase_ = re.compile(r"""<FILL\s+[^>]*>""") def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = code.split("\n" ) lowercase__ = 0 while idx < len(lowerCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCAmelCase__ ): return re.search(r"^(\s*)\S" , lines[idx] ).groups()[0] return "" def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = len(get_indent(lowerCAmelCase__ ) ) > 0 if has_indent: lowercase__ = f'''class Bla:\n{code}''' lowercase__ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowerCAmelCase__ ) lowercase__ = black.format_str(lowerCAmelCase__ , mode=lowerCAmelCase__ ) lowercase__ , lowercase__ = style_docstrings_in_code(lowerCAmelCase__ ) return result[len("class Bla:\n" ) :] if has_indent else result def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): with open(lowerCAmelCase__ , "r" , encoding="utf-8" , newline="\n" ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCAmelCase__ ): lowercase__ = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowercase__ , lowercase__ , lowercase__ = search.groups() lowercase__ = find_code_in_diffusers(lowerCAmelCase__ ) lowercase__ = get_indent(lowerCAmelCase__ ) lowercase__ = line_index + 1 if indent == theoretical_indent else line_index + 2 lowercase__ = theoretical_indent lowercase__ = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowercase__ = True while line_index < len(lowerCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(lowerCAmelCase__ ): break lowercase__ = lines[line_index] lowercase__ = _should_continue(lowerCAmelCase__ , lowerCAmelCase__ ) and re.search(f'''^{indent}# End copy''' , lowerCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase__ = lines[start_index:line_index] lowercase__ = "".join(lowerCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies lowercase__ = [line for line in theoretical_code.split("\n" ) if _re_copy_warning.search(lowerCAmelCase__ ) is None] lowercase__ = "\n".join(lowerCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCAmelCase__ ) > 0: lowercase__ = replace_pattern.replace("with" , "" ).split("," ) lowercase__ = [_re_replace_pattern.search(lowerCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue lowercase__ , lowercase__ , lowercase__ = pattern.groups() lowercase__ = re.sub(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if option.strip() == "all-casing": lowercase__ = re.sub(obja.lower() , obja.lower() , lowerCAmelCase__ ) lowercase__ = re.sub(obja.upper() , obja.upper() , lowerCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowercase__ = blackify(lines[start_index - 1] + theoretical_code ) lowercase__ = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowercase__ = lines[:start_index] + [theoretical_code] + lines[line_index:] lowercase__ = start_index + 1 if overwrite and len(lowerCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'''Detected changes, rewriting {filename}.''' ) with open(lowerCAmelCase__ , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lowerCAmelCase__ ) return diffs def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = False ): lowercase__ = glob.glob(os.path.join(lowerCAmelCase__ , "**/*.py" ) , recursive=lowerCAmelCase__ ) lowercase__ = [] for filename in all_files: lowercase__ = is_copy_consistent(lowerCAmelCase__ , lowerCAmelCase__ ) diffs += [f'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(lowerCAmelCase__ ) > 0: lowercase__ = "\n".join(lowerCAmelCase__ ) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them." ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") lowercase_ = parser.parse_args() check_copies(args.fix_and_overwrite)
705
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() lowercase_ = { """bart""": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), """bert""": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-base-cased-finetuned-mrpc""": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """dpr""": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), """gpt2""": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlnet""": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm""": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm-roberta""": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """transfo-xl""": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """openai-gpt""": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """roberta""": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """layoutlm""": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), """roberta-large-mnli""": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """camembert""": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """flaubert""": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert""": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert-base-distilled-squad""": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert""": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert-visual-feature-encoder""": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """ctrl""": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """albert""": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """t5""": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """electra""": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """wav2vec2""": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) lowercase__ = config_class.from_json_file(SCREAMING_SNAKE_CASE_ ) lowercase__ = True lowercase__ = True print(f'''Building TensorFlow model from configuration: {config}''' ) lowercase__ = model_class(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): lowercase__ = cached_file( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: lowercase__ = load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if compare_with_pt_model: lowercase__ = tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE_ ) # build the network lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" ) lowercase__ = pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE_ , config=SCREAMING_SNAKE_CASE_ , state_dict=SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): lowercase__ = pt_model(**pt_model.dummy_inputs ) lowercase__ = pto[0].numpy() lowercase__ = tfo[0].numpy() lowercase__ = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(SCREAMING_SNAKE_CASE_ , save_format="h5" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ): if args_model_type is None: lowercase__ = list(MODEL_CLASSES.keys() ) else: lowercase__ = [args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE_ , start=1 ): print("=" * 100 ) print(f''' Converting model type {j}/{len(SCREAMING_SNAKE_CASE_ )}: {model_type}''' ) print("=" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: lowercase__ = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: lowercase__ = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , start=1 ): print("-" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue lowercase__ = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE_ )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 100 ) if config_shortcut_name in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = config_shortcut_name if model_shortcut_name in aws_model_maps: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE_ ): lowercase__ = "converted_model" convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE_ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE_ , config_file=SCREAMING_SNAKE_CASE_ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE_ , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=SCREAMING_SNAKE_CASE_ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE_ ) os.remove(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_dump_path""", default=None, type=str, required=True, help="""Path to the output Tensorflow dump file.""" ) parser.add_argument( """--model_type""", default=None, type=str, help=( F'Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ' """convert all the models from AWS.""" ), ) parser.add_argument( """--pytorch_checkpoint_path""", default=None, type=str, help=( """Path to the PyTorch checkpoint path or shortcut name to download from AWS. """ """If not given, will download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--config_file""", default=None, type=str, help=( """The config json file corresponding to the pre-trained model. \n""" """This specifies the model architecture. If not given and """ """--pytorch_checkpoint_path is not given or is a shortcut name """ """use the configuration associated to the shortcut name on the AWS""" ), ) parser.add_argument( """--compare_with_pt_model""", action="""store_true""", help="""Compare Tensorflow and PyTorch model predictions.""" ) parser.add_argument( """--use_cached_models""", action="""store_true""", help="""Use cached models if possible instead of updating to latest checkpoint versions.""", ) parser.add_argument( """--remove_cached_files""", action="""store_true""", help="""Remove pytorch models after conversion (save memory when converting in batches).""", ) parser.add_argument("""--only_convert_finetuned_models""", action="""store_true""", help="""Only convert finetuned models.""") lowercase_ = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
37
0
from collections import namedtuple import requests from lxml import html # type: ignore lowercase_ = namedtuple("""covid_data""", """cases deaths recovered""") def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = "https://www.worldometers.info/coronavirus/" ): lowercase__ = "//div[@class = \"maincounter-number\"]/span/text()" return covid_data(*html.fromstring(requests.get(lowerCAmelCase_ ).content ).xpath(lowerCAmelCase_ ) ) lowercase_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
706
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
0
from __future__ import annotations from statistics import mean def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [0] * no_of_processes lowercase__ = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(_snake_case ): lowercase__ = burst_time[i] lowercase__ = [] lowercase__ = 0 lowercase__ = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: lowercase__ = [] lowercase__ = -1 for i in range(_snake_case ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(_snake_case ) if len(_snake_case ) > 0: lowercase__ = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: lowercase__ = i total_time += burst_time[target_process] completed += 1 lowercase__ = 0 lowercase__ = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = [0] * no_of_processes for i in range(_snake_case ): lowercase__ = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("""[TEST CASE 01]""") lowercase_ = 4 lowercase_ = [2, 5, 3, 7] lowercase_ = [0, 0, 0, 0] lowercase_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowercase_ = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("""PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time""") for i, process_id in enumerate(list(range(1, 5))): print( F'{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t' F'{waiting_time[i]}\t\t\t\t{turn_around_time[i]}' ) print(F'\nAverage waiting time = {mean(waiting_time):.5f}') print(F'Average turnaround time = {mean(turn_around_time):.5f}')
707
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : def __init__( self : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Union[str, Any], __lowercase : str, __lowercase : List[Any], __lowercase : List[str]=0.2, __lowercase : List[str]=0.2 ): lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = conva_get[:2] lowercase__ = conva_get[2] lowercase__ = size_pa lowercase__ = rate_w lowercase__ = rate_t lowercase__ = [ np.mat(-1 * np.random.rand(self.conva[0], self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self : Any, __lowercase : List[str] ): # save model dict with pickle lowercase__ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(__lowercase, "wb" ) as f: pickle.dump(__lowercase, __lowercase ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls : Dict, __lowercase : Union[str, Any] ): # read saved model with open(__lowercase, "rb" ) as f: lowercase__ = pickle.load(__lowercase ) # noqa: S301 lowercase__ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) lowercase__ = model_dic.get("size_pooling1" ) lowercase__ = model_dic.get("num_bp1" ) lowercase__ = model_dic.get("num_bp2" ) lowercase__ = model_dic.get("num_bp3" ) lowercase__ = model_dic.get("rate_weight" ) lowercase__ = model_dic.get("rate_thre" ) # create model instance lowercase__ = CNN(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) # modify model parameter lowercase__ = model_dic.get("w_conv1" ) lowercase__ = model_dic.get("wkj" ) lowercase__ = model_dic.get("vji" ) lowercase__ = model_dic.get("thre_conv1" ) lowercase__ = model_dic.get("thre_bp2" ) lowercase__ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self : str, __lowercase : List[Any] ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self : List[str], __lowercase : Optional[Any] ): return round(__lowercase, 3 ) def A__ ( self : Optional[Any], __lowercase : Dict, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : str ): # convolution process lowercase__ = convs[0] lowercase__ = convs[1] lowercase__ = np.shape(__lowercase )[0] # get the data slice of original image data, data_focus lowercase__ = [] for i_focus in range(0, size_data - size_conv + 1, __lowercase ): for j_focus in range(0, size_data - size_conv + 1, __lowercase ): lowercase__ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ = [] lowercase__ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__lowercase ): lowercase__ = [] for i_focus in range(len(__lowercase ) ): lowercase__ = ( np.sum(np.multiply(data_focus[i_focus], w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape( __lowercase, __lowercase ) data_featuremap.append(__lowercase ) # expanding the data slice to One dimenssion lowercase__ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__lowercase ) ) lowercase__ = np.asarray(__lowercase ) return focus_list, data_featuremap def A__ ( self : List[Any], __lowercase : Any, __lowercase : List[Any], __lowercase : Union[str, Any]="average_pool" ): # pooling process lowercase__ = len(featuremaps[0] ) lowercase__ = int(size_map / size_pooling ) lowercase__ = [] for i_map in range(len(__lowercase ) ): lowercase__ = featuremaps[i_map] lowercase__ = [] for i_focus in range(0, __lowercase, __lowercase ): for j_focus in range(0, __lowercase, __lowercase ): lowercase__ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape(__lowercase, __lowercase ) featuremap_pooled.append(__lowercase ) return featuremap_pooled def A__ ( self : str, __lowercase : Optional[Any] ): # expanding three dimension data to one dimension list lowercase__ = [] for i in range(len(__lowercase ) ): lowercase__ = np.shape(data[i] ) lowercase__ = data[i].reshape(1, shapes[0] * shapes[1] ) lowercase__ = data_listed.getA().tolist()[0] data_expanded.extend(__lowercase ) lowercase__ = np.asarray(__lowercase ) return data_expanded def A__ ( self : Optional[int], __lowercase : Optional[int] ): # expanding matrix to one dimension list lowercase__ = np.asarray(__lowercase ) lowercase__ = np.shape(__lowercase ) lowercase__ = data_mat.reshape(1, shapes[0] * shapes[1] ) return data_expanded def A__ ( self : str, __lowercase : Tuple, __lowercase : List[Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = [] lowercase__ = 0 for i_map in range(__lowercase ): lowercase__ = np.ones((size_map, size_map) ) for i in range(0, __lowercase, __lowercase ): for j in range(0, __lowercase, __lowercase ): lowercase__ = pd_pool[ i_pool ] lowercase__ = i_pool + 1 lowercase__ = np.multiply( __lowercase, np.multiply(out_map[i_map], (1 - out_map[i_map]) ) ) pd_all.append(__lowercase ) return pd_all def A__ ( self : Tuple, __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : List[str]=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(__lowercase )) ) print((" - - Shape: Teach_Data ", np.shape(__lowercase )) ) lowercase__ = 0 lowercase__ = [] lowercase__ = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase__ = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(__lowercase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ = np.asmatrix(datas_train[p] ) lowercase__ = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = np.shape(__lowercase ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = np.dot(__lowercase, self.vji.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = np.dot(__lowercase, self.wkj.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ = np.multiply( (data_teach - bp_outa), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.multiply( np.dot(__lowercase, self.wkj ), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.dot(__lowercase, self.vji ) lowercase__ = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ = pd_conva_pooled.T.getA().tolist() lowercase__ = self._calculate_gradient_from_pool( __lowercase, __lowercase, shape_featuremapa[0], shape_featuremapa[1], self.size_poolinga, ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ = self.rate_weight * np.dot(__lowercase, __lowercase ) lowercase__ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ = rp + 1 lowercase__ = error_count / patterns all_mse.append(__lowercase ) def draw_error(): lowercase__ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__lowercase, "+-" ) plt.plot(__lowercase, "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(__lowercase, alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self : List[str], __lowercase : Optional[int] ): # model predict lowercase__ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(__lowercase )) ) for p in range(len(__lowercase ) ): lowercase__ = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = bp_outa * self.vji.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = bp_outa * self.wkj.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ = [list(map(self.do_round, __lowercase ) ) for each in produce_out] return np.asarray(__lowercase ) def A__ ( self : int, __lowercase : Any ): # return the data of image after convoluting process so we can check it out lowercase__ = np.asmatrix(__lowercase ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
37
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all BART models at https://huggingface.co/models?filter=bart lowercase_ = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, 'tokenizer_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json', }, } lowercase_ = { 'facebook/bart-base': 1024, 'facebook/bart-large': 1024, 'facebook/bart-large-mnli': 1024, 'facebook/bart-large-cnn': 1024, 'facebook/bart-large-xsum': 1024, 'yjernite/bart_eli5': 1024, } class _snake_case ( lowerCAmelCase__): UpperCamelCase__ : Optional[int] =VOCAB_FILES_NAMES UpperCamelCase__ : str =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Any =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Any =["input_ids", "attention_mask"] UpperCamelCase__ : Union[str, Any] =BartTokenizer def __init__( self : Any, __lowercase : List[str]=None, __lowercase : Tuple=None, __lowercase : Tuple=None, __lowercase : Optional[int]="replace", __lowercase : List[Any]="<s>", __lowercase : Any="</s>", __lowercase : int="</s>", __lowercase : Dict="<s>", __lowercase : Tuple="<unk>", __lowercase : Optional[Any]="<pad>", __lowercase : Any="<mask>", __lowercase : Tuple=False, __lowercase : int=True, **__lowercase : Union[str, Any], ): super().__init__( _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE, tokenizer_file=_SCREAMING_SNAKE_CASE, errors=_SCREAMING_SNAKE_CASE, bos_token=_SCREAMING_SNAKE_CASE, eos_token=_SCREAMING_SNAKE_CASE, sep_token=_SCREAMING_SNAKE_CASE, cls_token=_SCREAMING_SNAKE_CASE, unk_token=_SCREAMING_SNAKE_CASE, pad_token=_SCREAMING_SNAKE_CASE, mask_token=_SCREAMING_SNAKE_CASE, add_prefix_space=_SCREAMING_SNAKE_CASE, trim_offsets=_SCREAMING_SNAKE_CASE, **_SCREAMING_SNAKE_CASE, ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space", _SCREAMING_SNAKE_CASE ) != add_prefix_space: lowercase__ = getattr(_SCREAMING_SNAKE_CASE, pre_tok_state.pop("type" ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**_SCREAMING_SNAKE_CASE ) lowercase__ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase__ = "post_processor" lowercase__ = getattr(self.backend_tokenizer, _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state["sep"] ) if "cls" in state: lowercase__ = tuple(state["cls"] ) lowercase__ = False if state.get("add_prefix_space", _SCREAMING_SNAKE_CASE ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get("trim_offsets", _SCREAMING_SNAKE_CASE ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(_SCREAMING_SNAKE_CASE, state.pop("type" ) ) lowercase__ = component_class(**_SCREAMING_SNAKE_CASE ) setattr(self.backend_tokenizer, _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE ) @property def A__ ( self : Union[str, Any] ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self : Any, __lowercase : List[str] ): lowercase__ = AddedToken(_SCREAMING_SNAKE_CASE, lstrip=_SCREAMING_SNAKE_CASE, rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE ) else value lowercase__ = value def A__ ( self : Optional[Any], *__lowercase : Tuple, **__lowercase : Tuple ): lowercase__ = kwargs.get("is_split_into_words", _SCREAMING_SNAKE_CASE ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*_SCREAMING_SNAKE_CASE, **_SCREAMING_SNAKE_CASE ) def A__ ( self : Tuple, *__lowercase : str, **__lowercase : Union[str, Any] ): lowercase__ = kwargs.get("is_split_into_words", _SCREAMING_SNAKE_CASE ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*_SCREAMING_SNAKE_CASE, **_SCREAMING_SNAKE_CASE ) def A__ ( self : Any, __lowercase : Optional[Any], __lowercase : List[Any] = None ): lowercase__ = self._tokenizer.model.save(_SCREAMING_SNAKE_CASE, name=_SCREAMING_SNAKE_CASE ) return tuple(_SCREAMING_SNAKE_CASE ) def A__ ( self : List[Any], __lowercase : Union[str, Any], __lowercase : Optional[Any]=None ): lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self : List[str], __lowercase : Dict, __lowercase : Optional[int] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
708
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=SCREAMING_SNAKE_CASE_ , num_labels=1000 , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): lowercase__ = get_config(SCREAMING_SNAKE_CASE_ ) # load original model from timm lowercase__ = create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=SCREAMING_SNAKE_CASE_ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) lowercase__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # verify logits with torch.no_grad(): lowercase__ = model(SCREAMING_SNAKE_CASE_ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(SCREAMING_SNAKE_CASE_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE_ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowercase_ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
37
0
import numpy as np def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 1e-12 , SCREAMING_SNAKE_CASE_ = 100 , ): assert np.shape(A_ )[0] == np.shape(A_ )[1] # Ensure proper dimensionality. assert np.shape(A_ )[0] == np.shape(A_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(A_ ) == np.iscomplexobj(A_ ) lowercase__ = np.iscomplexobj(A_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(A_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. lowercase__ = False lowercase__ = 0 lowercase__ = 0 lowercase__ = 1e12 while not convergence: # Multiple matrix by the vector. lowercase__ = np.dot(A_ , A_ ) # Normalize the resulting output vector. lowercase__ = w / np.linalg.norm(A_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) lowercase__ = vector.conj().T if is_complex else vector.T lowercase__ = np.dot(A_ , np.dot(A_ , A_ ) ) # Check convergence. lowercase__ = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: lowercase__ = True lowercase__ = lambda_ if is_complex: lowercase__ = np.real(lambda_ ) return lambda_, vector def __lowerCAmelCase ( ): lowercase__ = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) lowercase__ = np.array([41, 4, 20] ) lowercase__ = real_input_matrix.astype(np.complexaaa ) lowercase__ = np.triu(1J * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T lowercase__ = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": lowercase__ = real_input_matrix lowercase__ = real_vector elif problem_type == "complex": lowercase__ = complex_input_matrix lowercase__ = complex_vector # Our implementation. lowercase__ , lowercase__ = power_iteration(A_ , A_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). lowercase__ , lowercase__ = np.linalg.eigh(A_ ) # Last eigenvalue is the maximum one. lowercase__ = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. lowercase__ = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(A_ ) - np.abs(A_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
709
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _snake_case ( lowercase__): def __init__( self : Optional[Any], __lowercase : str = "▁", __lowercase : bool = True, __lowercase : Union[str, AddedToken] = "<unk>", __lowercase : Union[str, AddedToken] = "</s>", __lowercase : Union[str, AddedToken] = "<pad>", ): lowercase__ = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } lowercase__ = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowercase__ = token_dict["token"] lowercase__ = Tokenizer(Unigram() ) lowercase__ = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ), " " ), normalizers.Lowercase(), ] ) lowercase__ = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) lowercase__ = decoders.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ) lowercase__ = TemplateProcessing( single=F'''$A {self.special_tokens["eos"]["token"]}''', special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])], ) lowercase__ = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase, __lowercase ) def A__ ( self : Union[str, Any], __lowercase : Union[str, List[str]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) if isinstance(__lowercase, __lowercase ): lowercase__ = [files] self._tokenizer.train(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : List[Any], __lowercase : Union[Iterator[str], Iterator[Iterator[str]]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) self._tokenizer.train_from_iterator(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : str ): lowercase__ = json.loads(self._tokenizer.to_str() ) lowercase__ = self.special_tokens["unk"]["id"] lowercase__ = Tokenizer.from_str(json.dumps(__lowercase ) )
37
0
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def A__ ( *__lowercase : Dict, **__lowercase : Tuple ): pass @is_pipeline_test @require_vision class _snake_case ( unittest.TestCase): @require_torch def A__ ( self : str ): lowercase__ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification", ) lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(_A, candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(_A ), [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["A", "B", "C"], batch_size=2 ) self.assertEqual( nested_simplify(_A ), [ [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], ], ) @require_tf def A__ ( self : Optional[int] ): lowercase__ = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification", framework="tf" ) lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(_A, candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(_A ), [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["A", "B", "C"], batch_size=2 ) self.assertEqual( nested_simplify(_A ), [ [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], [ {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, {"score": 0.333, "label": ANY(_A )}, ], ], ) @slow @require_torch def A__ ( self : Union[str, Any] ): lowercase__ = pipeline( task="zero-shot-image-classification", model="openai/clip-vit-base-patch32", ) # This is an image of 2 cats with remotes and no planes lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(_A, candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(_A ), [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["cat", "plane", "remote"], batch_size=2 ) self.assertEqual( nested_simplify(_A ), [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5, ) @slow @require_tf def A__ ( self : Optional[int] ): lowercase__ = pipeline( task="zero-shot-image-classification", model="openai/clip-vit-base-patch32", framework="tf" ) # This is an image of 2 cats with remotes and no planes lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowercase__ = image_classifier(_A, candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(_A ), [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ) lowercase__ = image_classifier([image] * 5, candidate_labels=["cat", "plane", "remote"], batch_size=2 ) self.assertEqual( nested_simplify(_A ), [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5, )
710
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase__ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowercase__ = f'''{src_lang}-{tgt_lang}''' lowercase__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) lowercase__ = os.path.join(SCREAMING_SNAKE_CASE_ , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
37
0
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class _snake_case : UpperCamelCase__ : List[str] =BlenderbotConfig UpperCamelCase__ : Dict ={} UpperCamelCase__ : Optional[int] ="""gelu""" def __init__( self : Any, __lowercase : Dict, __lowercase : int=13, __lowercase : Union[str, Any]=7, __lowercase : str=True, __lowercase : Any=False, __lowercase : List[str]=99, __lowercase : Optional[Any]=32, __lowercase : Union[str, Any]=2, __lowercase : Any=4, __lowercase : str=37, __lowercase : Any=0.1, __lowercase : str=0.1, __lowercase : Tuple=20, __lowercase : Tuple=2, __lowercase : Optional[int]=1, __lowercase : Tuple=0, ): lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = eos_token_id lowercase__ = pad_token_id lowercase__ = bos_token_id def A__ ( self : List[Any] ): lowercase__ = ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ) lowercase__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ), 1 ) lowercase__ = tf.concat([input_ids, eos_tensor], axis=1 ) lowercase__ = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) lowercase__ = self.config_cls( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_ids=[2], bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.pad_token_id, **self.config_updates, ) lowercase__ = prepare_blenderbot_inputs_dict(__a, __a, __a ) return config, inputs_dict def A__ ( self : Optional[int], __lowercase : Any, __lowercase : int ): lowercase__ = TFBlenderbotModel(config=__a ).get_decoder() lowercase__ = inputs_dict["input_ids"] lowercase__ = input_ids[:1, :] lowercase__ = inputs_dict["attention_mask"][:1, :] lowercase__ = inputs_dict["head_mask"] lowercase__ = 1 # first forward pass lowercase__ = model(__a, attention_mask=__a, head_mask=__a, use_cache=__a ) lowercase__ , lowercase__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase__ = ids_tensor((self.batch_size, 3), config.vocab_size ) lowercase__ = tf.cast(ids_tensor((self.batch_size, 3), 2 ), tf.inta ) # append to next input_ids and lowercase__ = tf.concat([input_ids, next_tokens], axis=-1 ) lowercase__ = tf.concat([attention_mask, next_attn_mask], axis=-1 ) lowercase__ = model(__a, attention_mask=__a )[0] lowercase__ = model(__a, attention_mask=__a, past_key_values=__a )[0] self.parent.assertEqual(next_tokens.shape[1], output_from_past.shape[1] ) # select random slice lowercase__ = int(ids_tensor((1,), output_from_past.shape[-1] ) ) lowercase__ = output_from_no_past[:, -3:, random_slice_idx] lowercase__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__a, __a, rtol=1e-3 ) def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , ): if attention_mask is None: lowercase__ = tf.cast(tf.math.not_equal(__snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowercase__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowercase__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowercase__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowercase__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _snake_case ( lowercase__ , lowercase__ , unittest.TestCase): UpperCamelCase__ : Tuple =(TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () UpperCamelCase__ : List[Any] =(TFBlenderbotForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ : Optional[Any] =( { """conversational""": TFBlenderbotForConditionalGeneration, """feature-extraction""": TFBlenderbotModel, """summarization""": TFBlenderbotForConditionalGeneration, """text2text-generation""": TFBlenderbotForConditionalGeneration, """translation""": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ : str =True UpperCamelCase__ : Union[str, Any] =False UpperCamelCase__ : str =False def A__ ( self : Union[str, Any] ): lowercase__ = TFBlenderbotModelTester(self ) lowercase__ = ConfigTester(self, config_class=__a ) def A__ ( self : Union[str, Any] ): self.config_tester.run_common_tests() def A__ ( self : Optional[Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a ) @require_tokenizers @require_tf class _snake_case ( unittest.TestCase): UpperCamelCase__ : int =["""My friends are cool but they eat too many carbs."""] UpperCamelCase__ : Optional[int] ="""facebook/blenderbot-400M-distill""" @cached_property def A__ ( self : List[Any] ): return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def A__ ( self : int ): lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def A__ ( self : str ): lowercase__ = self.tokenizer(self.src_text, return_tensors="tf" ) lowercase__ = self.model.generate( model_inputs.input_ids, ) lowercase__ = self.tokenizer.batch_decode(generated_ids.numpy(), skip_special_tokens=__a )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
711
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =TransfoXLTokenizer UpperCamelCase__ : List[Any] =False UpperCamelCase__ : List[Any] =False def A__ ( self : Union[str, Any] ): super().setUp() lowercase__ = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : Union[str, Any], **__lowercase : Any ): lowercase__ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname, **__lowercase ) def A__ ( self : Tuple, __lowercase : Optional[int] ): lowercase__ = "<unk> UNwanted , running" lowercase__ = "<unk> unwanted, running" return input_text, output_text def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(vocab_file=self.vocab_file, lower_case=__lowercase ) lowercase__ = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowercase, ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ), [0, 4, 8, 7] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["hello", "!", "how", "are", "you", "?"] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) lowercase__ = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" lowercase__ = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowercase ), __lowercase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowercase ), __lowercase ) def A__ ( self : List[str] ): lowercase__ = self.get_tokenizer() lowercase__ = len(__lowercase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1", 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowercase ), original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ), [1] ) self.assertEqual(tokenizer.decode([1] ), "new1" )
37
0
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = filter(lambda SCREAMING_SNAKE_CASE_ : p.requires_grad , model.parameters() ) lowercase__ = sum([np.prod(p.size() ) for p in model_parameters] ) return params lowercase_ = logging.getLogger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if metric == "rouge2": lowercase__ = """{val_avg_rouge2:.4f}-{step_count}""" elif metric == "bleu": lowercase__ = """{val_avg_bleu:.4f}-{step_count}""" elif metric == "em": lowercase__ = """{val_avg_em:.4f}-{step_count}""" else: raise NotImplementedError( f'''seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this''' " function." ) lowercase__ = ModelCheckpoint( dirpath=lowerCamelCase_ , filename=lowerCamelCase_ , monitor=f'''val_{metric}''' , mode="max" , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return EarlyStopping( monitor=f'''val_{metric}''' , mode="min" if "loss" in metric else "max" , patience=lowerCamelCase_ , verbose=lowerCamelCase_ , ) class _snake_case ( pl.Callback): def A__ ( self : Any, __lowercase : List[Any], __lowercase : str ): lowercase__ = {F'''lr_group_{i}''': param["""lr"""] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCamelCase_ ) @rank_zero_only def A__ ( self : List[Any], __lowercase : pl.Trainer, __lowercase : pl.LightningModule, __lowercase : str, __lowercase : List[str]=True ): logger.info(F'''***** {type_path} results at step {trainer.global_step:05d} *****''' ) lowercase__ = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results lowercase__ = Path(pl_module.hparams.output_dir ) if type_path == "test": lowercase__ = od / """test_results.txt""" lowercase__ = od / """test_generations.txt""" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. lowercase__ = od / F'''{type_path}_results/{trainer.global_step:05d}.txt''' lowercase__ = od / F'''{type_path}_generations/{trainer.global_step:05d}.txt''' results_file.parent.mkdir(exist_ok=lowerCamelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCamelCase_ ) with open(lowerCamelCase_, "a+" ) as writer: for key in sorted(lowerCamelCase_ ): if key in ["log", "progress_bar", "preds"]: continue lowercase__ = metrics[key] if isinstance(lowerCamelCase_, torch.Tensor ): lowercase__ = val.item() lowercase__ = F'''{key}: {val:.6f}\n''' writer.write(lowerCamelCase_ ) if not save_generations: return if "preds" in metrics: lowercase__ = """\n""".join(metrics["preds"] ) generations_file.open("w+" ).write(lowerCamelCase_ ) @rank_zero_only def A__ ( self : str, __lowercase : Optional[int], __lowercase : Any ): try: lowercase__ = pl_module.model.model.num_parameters() except AttributeError: lowercase__ = pl_module.model.num_parameters() lowercase__ = count_trainable_parameters(lowerCamelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1e6, "grad_mp": n_trainable_pars / 1e6} ) @rank_zero_only def A__ ( self : List[str], __lowercase : pl.Trainer, __lowercase : pl.LightningModule ): save_json(pl_module.metrics, pl_module.metrics_save_path ) return self._write_logs(lowerCamelCase_, lowerCamelCase_, "test" ) @rank_zero_only def A__ ( self : List[str], __lowercase : pl.Trainer, __lowercase : Any ): save_json(pl_module.metrics, pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
712
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser(SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.parse_args_into_dataclasses()[0] lowercase__ = TensorFlowBenchmark(args=SCREAMING_SNAKE_CASE_ ) try: lowercase__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowercase__ = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowercase__ = " ".join(str(SCREAMING_SNAKE_CASE_ ).split(" " )[:-1] ) lowercase__ = "" lowercase__ = eval(str(SCREAMING_SNAKE_CASE_ ).split(" " )[-1] ) lowercase__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = full_error_msg + begin_error_msg + str(SCREAMING_SNAKE_CASE_ ) raise ValueError(SCREAMING_SNAKE_CASE_ ) benchmark.run() if __name__ == "__main__": main()
37
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase_ = { """configuration_groupvit""": [ """GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GroupViTConfig""", """GroupViTOnnxConfig""", """GroupViTTextConfig""", """GroupViTVisionConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GroupViTModel""", """GroupViTPreTrainedModel""", """GroupViTTextModel""", """GroupViTVisionModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFGroupViTModel""", """TFGroupViTPreTrainedModel""", """TFGroupViTTextModel""", """TFGroupViTVisionModel""", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
713
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowercase_ = """<<<<<<< This should probably be modified because it mentions: """ lowercase_ = """======= >>>>>>> """ lowercase_ = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] lowercase_ = [ # (pattern, replacement) # Order is important here for some replacements (r"""tfds\.core""", r"""datasets"""), (r"""tf\.io\.gfile\.GFile""", r"""open"""), (r"""tf\.([\w\d]+)""", r"""datasets.Value('\1')"""), (r"""tfds\.features\.Text\(\)""", r"""datasets.Value('string')"""), (r"""tfds\.features\.Text\(""", r"""datasets.Value('string'),"""), (r"""features\s*=\s*tfds.features.FeaturesDict\(""", r"""features=datasets.Features("""), (r"""tfds\.features\.FeaturesDict\(""", r"""dict("""), (r"""The TensorFlow Datasets Authors""", r"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (r"""tfds\.""", r"""datasets."""), (r"""dl_manager\.manual_dir""", r"""self.config.data_dir"""), (r"""self\.builder_config""", r"""self.config"""), ] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return ConvertCommand(args.tfds_path , args.datasets_directory ) class _snake_case ( lowercase__): @staticmethod def A__ ( __lowercase : ArgumentParser ): lowercase__ = parser.add_parser( "convert", help="Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.", ) train_parser.add_argument( "--tfds_path", type=__lowercase, required=__lowercase, help="Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.", ) train_parser.add_argument( "--datasets_directory", type=__lowercase, required=__lowercase, help="Path to the HuggingFace Datasets folder." ) train_parser.set_defaults(func=__lowercase ) def __init__( self : Tuple, __lowercase : str, __lowercase : str, *__lowercase : Tuple ): lowercase__ = get_logger("datasets-cli/converting" ) lowercase__ = tfds_path lowercase__ = datasets_directory def A__ ( self : Any ): if os.path.isdir(self._tfds_path ): lowercase__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__ = os.path.dirname(self._tfds_path ) else: raise ValueError("--tfds_path is neither a directory nor a file. Please check path." ) lowercase__ = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) lowercase__ = [] lowercase__ = [] lowercase__ = {} if os.path.isdir(self._tfds_path ): lowercase__ = os.listdir(__lowercase ) else: lowercase__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) if not os.path.isfile(__lowercase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("Skipping file" ) continue with open(__lowercase, encoding="utf-8" ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = False lowercase__ = False lowercase__ = [] for line in lines: lowercase__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__ = "import datasets\n" elif "import tensorflow" in out_line: # order is important here lowercase__ = "" continue elif "from absl import logging" in out_line: lowercase__ = "from datasets import logging\n" elif "getLogger" in out_line: lowercase__ = out_line.replace("getLogger", "get_logger" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__ = True lowercase__ = list(filter(lambda __lowercase : e in out_line, __lowercase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowercase ) + "\n" ) out_lines.append(__lowercase ) out_lines.append(__lowercase ) continue else: for pattern, replacement in TO_CONVERT: lowercase__ = re.sub(__lowercase, __lowercase, __lowercase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__ = re.match(R"from\stensorflow_datasets.*import\s([^\.\r\n]+)", __lowercase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split("," ) ) lowercase__ = "from . import " + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__ = True out_lines.append(__lowercase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__ = f_name.replace(".py", "" ) lowercase__ = os.path.join(__lowercase, __lowercase ) lowercase__ = os.path.join(__lowercase, __lowercase ) os.makedirs(__lowercase, exist_ok=__lowercase ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowercase ) if needs_manual_update: with_manual_update.append(__lowercase ) with open(__lowercase, "w", encoding="utf-8" ) as f: f.writelines(__lowercase ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: lowercase__ = os.path.basename(__lowercase ) lowercase__ = imports_to_builder_map[f_name.replace(".py", "" )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowercase, __lowercase ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
37
0
from random import shuffle import tensorflow as tf from numpy import array def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = int(lowerCamelCase__ ) assert noofclusters < len(lowerCamelCase__ ) # Find out the dimensionality lowercase__ = len(vectors[0] ) # Will help select random centroids from among the available vectors lowercase__ = list(range(len(lowerCamelCase__ ) ) ) shuffle(lowerCamelCase__ ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. lowercase__ = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION lowercase__ = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points lowercase__ = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(lowerCamelCase__ ) ] ##These nodes will assign the centroid Variables the appropriate ##values lowercase__ = tf.placeholder("float64" , [dim] ) lowercase__ = [] for centroid in centroids: cent_assigns.append(tf.assign(lowerCamelCase__ , lowerCamelCase__ ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) lowercase__ = [tf.Variable(0 ) for i in range(len(lowerCamelCase__ ) )] ##These nodes will assign an assignment Variable the appropriate ##value lowercase__ = tf.placeholder("int32" ) lowercase__ = [] for assignment in assignments: cluster_assigns.append(tf.assign(lowerCamelCase__ , lowerCamelCase__ ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input lowercase__ = tf.placeholder("float" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors lowercase__ = tf.reduce_mean(lowerCamelCase__ , 0 ) ##Node for computing Euclidean distances # Placeholders for input lowercase__ = tf.placeholder("float" , [dim] ) lowercase__ = tf.placeholder("float" , [dim] ) lowercase__ = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(lowerCamelCase__ , lowerCamelCase__ ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input lowercase__ = tf.placeholder("float" , [noofclusters] ) lowercase__ = tf.argmin(lowerCamelCase__ , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. lowercase__ = tf.initialize_all_variables() # Initialize all variables sess.run(lowerCamelCase__ ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. lowercase__ = 100 for _ in range(lowerCamelCase__ ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(lowerCamelCase__ ) ): lowercase__ = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. lowercase__ = [ sess.run(lowerCamelCase__ , feed_dict={va: vect, va: sess.run(lowerCamelCase__ )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input lowercase__ = sess.run( lowerCamelCase__ , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(lowerCamelCase__ ): # Collect all the vectors assigned to this cluster lowercase__ = [ vectors[i] for i in range(len(lowerCamelCase__ ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location lowercase__ = sess.run( lowerCamelCase__ , feed_dict={mean_input: array(lowerCamelCase__ )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments lowercase__ = sess.run(lowerCamelCase__ ) lowercase__ = sess.run(lowerCamelCase__ ) return centroids, assignments
714
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase_ = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } lowercase_ = { """allenai/led-base-16384""": 1_6384, } class _snake_case ( lowercase__): UpperCamelCase__ : int =VOCAB_FILES_NAMES UpperCamelCase__ : Any =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =LEDTokenizer UpperCamelCase__ : Tuple =["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any], __lowercase : Optional[Any]=None, __lowercase : Dict=None, __lowercase : Tuple=None, __lowercase : Union[str, Any]="replace", __lowercase : Tuple="<s>", __lowercase : Optional[Any]="</s>", __lowercase : Tuple="</s>", __lowercase : List[str]="<s>", __lowercase : Tuple="<unk>", __lowercase : Dict="<pad>", __lowercase : Dict="<mask>", __lowercase : Any=False, __lowercase : Any=True, **__lowercase : List[Any], ): super().__init__( __lowercase, __lowercase, tokenizer_file=__lowercase, errors=__lowercase, bos_token=__lowercase, eos_token=__lowercase, sep_token=__lowercase, cls_token=__lowercase, unk_token=__lowercase, pad_token=__lowercase, mask_token=__lowercase, add_prefix_space=__lowercase, trim_offsets=__lowercase, **__lowercase, ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = getattr(__lowercase, pre_tok_state.pop("type" ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**__lowercase ) lowercase__ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase__ = "post_processor" lowercase__ = getattr(self.backend_tokenizer, __lowercase, __lowercase ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state["sep"] ) if "cls" in state: lowercase__ = tuple(state["cls"] ) lowercase__ = False if state.get("add_prefix_space", __lowercase ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get("trim_offsets", __lowercase ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(__lowercase, state.pop("type" ) ) lowercase__ = component_class(**__lowercase ) setattr(self.backend_tokenizer, __lowercase, __lowercase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A__ ( self : str ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self : Optional[int], __lowercase : Dict ): lowercase__ = AddedToken(__lowercase, lstrip=__lowercase, rstrip=__lowercase ) if isinstance(__lowercase, __lowercase ) else value lowercase__ = value def A__ ( self : Any, *__lowercase : List[Any], **__lowercase : Optional[Any] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowercase, **__lowercase ) def A__ ( self : int, *__lowercase : Union[str, Any], **__lowercase : List[str] ): lowercase__ = kwargs.get("is_split_into_words", __lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowercase, **__lowercase ) def A__ ( self : Optional[Any], __lowercase : str, __lowercase : Optional[str] = None ): lowercase__ = self._tokenizer.model.save(__lowercase, name=__lowercase ) return tuple(__lowercase ) def A__ ( self : List[str], __lowercase : int, __lowercase : Optional[int]=None ): lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self : int, __lowercase : List[int], __lowercase : Optional[List[int]] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self : Union[str, Any], __lowercase : Union[Dict[str, EncodedInput], BatchEncoding], __lowercase : Optional[int] = None, __lowercase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD, __lowercase : Optional[int] = None, __lowercase : Optional[bool] = None, ): lowercase__ = super()._pad( encoded_inputs=__lowercase, max_length=__lowercase, padding_strategy=__lowercase, pad_to_multiple_of=__lowercase, return_attention_mask=__lowercase, ) # Load from model defaults if return_attention_mask is None: lowercase__ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ = len(encoded_inputs["global_attention_mask"] ) != len(__lowercase ) if needs_to_be_padded: lowercase__ = len(__lowercase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
37
0
import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} # See all BART models at https://huggingface.co/models?filter=bart lowercase_ = { """vocab_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/vocab.json""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/vocab.json""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json""", }, """merges_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/merges.txt""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/merges.txt""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt""", }, } lowercase_ = { """facebook/bart-base""": 1024, """facebook/bart-large""": 1024, """facebook/bart-large-mnli""": 1024, """facebook/bart-large-cnn""": 1024, """facebook/bart-large-xsum""": 1024, """yjernite/bart_eli5""": 1024, } @lru_cache() def __lowerCAmelCase ( ): lowercase__ = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) lowercase__ = bs[:] lowercase__ = 0 for b in range(2**8 ): if b not in bs: bs.append(a__ ) cs.append(2**8 + n ) n += 1 lowercase__ = [chr(a__ ) for n in cs] return dict(zip(a__ , a__ ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = set() lowercase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ = char return pairs class _snake_case ( lowercase__): UpperCamelCase__ : str =VOCAB_FILES_NAMES UpperCamelCase__ : List[Any] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Optional[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Optional[Any] =["""input_ids""", """attention_mask"""] def __init__( self : int, __lowercase : List[Any], __lowercase : Union[str, Any], __lowercase : int="replace", __lowercase : Union[str, Any]="<s>", __lowercase : List[Any]="</s>", __lowercase : Union[str, Any]="</s>", __lowercase : Union[str, Any]="<s>", __lowercase : Optional[int]="<unk>", __lowercase : Optional[int]="<pad>", __lowercase : str="<mask>", __lowercase : Tuple=False, **__lowercase : List[str], ): lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else bos_token lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else eos_token lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else sep_token lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else cls_token lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else unk_token lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ = AddedToken(lowercase_, lstrip=lowercase_, rstrip=lowercase_ ) if isinstance(lowercase_, lowercase_ ) else mask_token super().__init__( errors=lowercase_, bos_token=lowercase_, eos_token=lowercase_, unk_token=lowercase_, sep_token=lowercase_, cls_token=lowercase_, pad_token=lowercase_, mask_token=lowercase_, add_prefix_space=lowercase_, **lowercase_, ) with open(lowercase_, encoding="utf-8" ) as vocab_handle: lowercase__ = json.load(lowercase_ ) lowercase__ = {v: k for k, v in self.encoder.items()} lowercase__ = errors # how to handle errors in decoding lowercase__ = bytes_to_unicode() lowercase__ = {v: k for k, v in self.byte_encoder.items()} with open(lowercase_, encoding="utf-8" ) as merges_handle: lowercase__ = merges_handle.read().split("\n" )[1:-1] lowercase__ = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ = dict(zip(lowercase_, range(len(lowercase_ ) ) ) ) lowercase__ = {} lowercase__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def A__ ( self : Any ): return len(self.encoder ) def A__ ( self : str ): return dict(self.encoder, **self.added_tokens_encoder ) def A__ ( self : List[str], __lowercase : str ): if token in self.cache: return self.cache[token] lowercase__ = tuple(lowercase_ ) lowercase__ = get_pairs(lowercase_ ) if not pairs: return token while True: lowercase__ = min(lowercase_, key=lambda __lowercase : self.bpe_ranks.get(lowercase_, float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ = bigram lowercase__ = [] lowercase__ = 0 while i < len(lowercase_ ): try: lowercase__ = word.index(lowercase_, lowercase_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ = j if word[i] == first and i < len(lowercase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ = tuple(lowercase_ ) lowercase__ = new_word if len(lowercase_ ) == 1: break else: lowercase__ = get_pairs(lowercase_ ) lowercase__ = " ".join(lowercase_ ) lowercase__ = word return word def A__ ( self : int, __lowercase : Optional[Any] ): lowercase__ = [] for token in re.findall(self.pat, lowercase_ ): lowercase__ = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowercase_ ).split(" " ) ) return bpe_tokens def A__ ( self : int, __lowercase : Optional[int] ): return self.encoder.get(lowercase_, self.encoder.get(self.unk_token ) ) def A__ ( self : Tuple, __lowercase : Optional[Any] ): return self.decoder.get(lowercase_ ) def A__ ( self : Union[str, Any], __lowercase : str ): lowercase__ = "".join(lowercase_ ) lowercase__ = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8", errors=self.errors ) return text def A__ ( self : List[Any], __lowercase : str, __lowercase : Optional[str] = None ): if not os.path.isdir(lowercase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( lowercase_, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ = os.path.join( lowercase_, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowercase_, "w", encoding="utf-8" ) as f: f.write(json.dumps(self.encoder, indent=2, sort_keys=lowercase_, ensure_ascii=lowercase_ ) + "\n" ) lowercase__ = 0 with open(lowercase_, "w", encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items(), key=lambda __lowercase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowercase__ = token_index writer.write(" ".join(lowercase_ ) + "\n" ) index += 1 return vocab_file, merge_file def A__ ( self : Union[str, Any], __lowercase : List[int], __lowercase : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ = [self.cls_token_id] lowercase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self : List[str], __lowercase : List[int], __lowercase : Optional[List[int]] = None, __lowercase : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_, token_ids_a=lowercase_, already_has_special_tokens=lowercase_ ) if token_ids_a is None: return [1] + ([0] * len(lowercase_ )) + [1] return [1] + ([0] * len(lowercase_ )) + [1, 1] + ([0] * len(lowercase_ )) + [1] def A__ ( self : Optional[Any], __lowercase : List[int], __lowercase : Optional[List[int]] = None ): lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self : Dict, __lowercase : List[str], __lowercase : int=False, **__lowercase : str ): lowercase__ = kwargs.pop("add_prefix_space", self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowercase_ ) > 0 and not text[0].isspace()): lowercase__ = " " + text return (text, kwargs)
715
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __lowerCAmelCase ( ): lowercase__ = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=SCREAMING_SNAKE_CASE_ ) env_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) launch_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) tpu_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) test_command_parser(subparsers=SCREAMING_SNAKE_CASE_ ) # Let's go lowercase__ = parser.parse_args() if not hasattr(SCREAMING_SNAKE_CASE_ , "func" ): parser.print_help() exit(1 ) # Run args.func(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
37
0
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = torch.load(snake_case__ , map_location="cpu" ) if "model" in sd.keys(): lowercase__ = torch.load(snake_case__ , map_location="cpu" )["model"] # pop unnecessary weights lowercase__ = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(snake_case__ ) lowercase__ = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: lowercase__ = sd.pop(snake_case__ ) lowercase__ = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: lowercase__ = sd[key] # We split QKV in separate Q,K,V lowercase__ = key.replace(".qkv_proj." , ".q_proj." ) lowercase__ = key.replace(".qkv_proj." , ".k_proj." ) lowercase__ = key.replace(".qkv_proj." , ".v_proj." ) lowercase__ = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 lowercase__ , lowercase__ , lowercase__ = torch.split(snake_case__ , depth // 3 , dim=0 ) lowercase__ = q lowercase__ = k lowercase__ = v del sd[key] return sd @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ): lowercase__ = load_checkpoint(snake_case__ ) if config is not None: lowercase__ = OPTConfig.from_pretrained(snake_case__ ) else: lowercase__ = OPTConfig() lowercase__ = OPTModel(snake_case__ ).half().eval() model.load_state_dict(snake_case__ ) # Check results Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--fairseq_path""", type=str, help=( """path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:""" """ https://huggingface.co/models?other=opt_metasq""" ), ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--hf_config""", default=None, type=str, help="""Define HF config.""") lowercase_ = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
716
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : Dict, __lowercase : int, __lowercase : Union[str, Any]=7, __lowercase : Union[str, Any]=3, __lowercase : Any=18, __lowercase : Union[str, Any]=30, __lowercase : Any=400, __lowercase : List[str]=True, __lowercase : Dict=None, __lowercase : List[str]=True, __lowercase : int=False, __lowercase : Union[str, Any]=True, __lowercase : str=True, __lowercase : Optional[int]=[0.5, 0.5, 0.5], __lowercase : List[Any]=[0.5, 0.5, 0.5], ): lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = do_resize lowercase__ = size if size is not None else {"height": 18, "width": 20} lowercase__ = do_thumbnail lowercase__ = do_align_axis lowercase__ = do_pad lowercase__ = do_normalize lowercase__ = image_mean lowercase__ = image_std def A__ ( self : Optional[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =DonutImageProcessor if is_vision_available() else None def A__ ( self : str ): lowercase__ = DonutImageProcessingTester(self ) @property def A__ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_resize" ) ) self.assertTrue(hasattr(__lowercase, "size" ) ) self.assertTrue(hasattr(__lowercase, "do_thumbnail" ) ) self.assertTrue(hasattr(__lowercase, "do_align_long_axis" ) ) self.assertTrue(hasattr(__lowercase, "do_pad" ) ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "image_mean" ) ) self.assertTrue(hasattr(__lowercase, "image_std" ) ) def A__ ( self : str ): lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {"height": 18, "width": 20} ) lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {"height": 42, "width": 42} ) # Previous config had dimensions in (width, height) order lowercase__ = self.image_processing_class.from_dict(self.image_processor_dict, size=(42, 84) ) self.assertEqual(image_processor.size, {"height": 84, "width": 42} ) def A__ ( self : List[str] ): pass @is_flaky() def A__ ( self : Dict ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Optional[Any] ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) @is_flaky() def A__ ( self : Tuple ): # Initialize image_processing lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched lowercase__ = image_processing(__lowercase, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), )
37
0
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging lowercase_ = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if "xprophetnet" in prophetnet_checkpoint_path: lowercase__ = XLMProphetNetForConditionalGenerationOld.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowercase__ = XLMProphetNetForConditionalGeneration.from_pretrained( SCREAMING_SNAKE_CASE_ , output_loading_info=SCREAMING_SNAKE_CASE_ ) else: lowercase__ = ProphetNetForConditionalGenerationOld.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowercase__ = ProphetNetForConditionalGeneration.from_pretrained( SCREAMING_SNAKE_CASE_ , output_loading_info=SCREAMING_SNAKE_CASE_ ) lowercase__ = ['''key_proj''', '''value_proj''', '''query_proj'''] lowercase__ = { '''self_attn''': '''ngram_self_attn''', '''cross_attn''': '''encoder_attn''', '''cross_attn_layer_norm''': '''encoder_attn_layer_norm''', '''feed_forward_layer_norm''': '''final_layer_norm''', '''feed_forward''': '''''', '''intermediate''': '''fc1''', '''output''': '''fc2''', '''key_proj''': '''k_proj''', '''query_proj''': '''q_proj''', '''value_proj''': '''v_proj''', '''word_embeddings''': '''embed_tokens''', '''embeddings_layer_norm''': '''emb_layer_norm''', '''relative_pos_embeddings''': '''relative_linear''', '''ngram_embeddings''': '''ngram_input_embed''', '''position_embeddings''': '''embed_positions''', } for key in loading_info["missing_keys"]: lowercase__ = key.split("." ) if attributes[0] == "lm_head": lowercase__ = prophet lowercase__ = prophet_old else: lowercase__ = prophet.prophetnet lowercase__ = prophet_old.model lowercase__ = False for attribute in attributes: if attribute in mapping: lowercase__ = mapping[attribute] if not hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = attribute elif hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" lowercase__ = old_model.weight logger.info(f'''{attribute} is initialized.''' ) lowercase__ = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" lowercase__ = old_model.bias logger.info(f'''{attribute} is initialized''' ) lowercase__ = True break elif attribute in special_keys and hasattr(SCREAMING_SNAKE_CASE_ , "in_proj_weight" ): lowercase__ = old_model.in_proj_weight.shape[0] // 3 lowercase__ = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": lowercase__ = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) lowercase__ = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": lowercase__ = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) lowercase__ = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": lowercase__ = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) lowercase__ = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) lowercase__ = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." lowercase__ = nn.Parameter(old_model.embed_positions.weight[:512, :] ) lowercase__ = True break if attribute.isdigit(): lowercase__ = model[int(SCREAMING_SNAKE_CASE_ )] lowercase__ = old_model[int(SCREAMING_SNAKE_CASE_ )] else: lowercase__ = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if old_attribute == "": lowercase__ = old_model else: if not hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'''{old_model} does not have {old_attribute}''' ) lowercase__ = getattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if not is_key_init: raise ValueError(f'''{key} was not correctly initialized!''' ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) prophet.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase_ = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
717
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _snake_case ( lowercase__): def A__ ( self : Optional[Any], __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) lowercase__ = input_file.read() lowercase__ = regexp.search(__lowercase ) return match def A__ ( self : str, __lowercase : str ): with open(__lowercase, encoding="utf-8" ) as input_file: lowercase__ = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()", re.DOTALL ) lowercase__ = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowercase__ = regexp.finditer(__lowercase ) lowercase__ = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(F'''open(...) must use utf-8 encoding in {dataset}''' ) def A__ ( self : Union[str, Any] ): lowercase__ = Path("./datasets" ) lowercase__ = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(F'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
37
0
lowercase_ = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowercase_ = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowercase_ = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): assert len(str(SCREAMING_SNAKE_CASE_ ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: lowercase__ = year // 100 lowercase__ = (5 * (century % 4) + 2) % 7 lowercase__ = year % 100 lowercase__ = centurian % 12 lowercase__ = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 lowercase__ = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) lowercase__ = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
718
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = 50 ): lowercase__ = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F'{solution() = }')
719
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowercase_ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _snake_case ( unittest.TestCase): def __init__( self : List[Any], __lowercase : int, __lowercase : Optional[int]=7, __lowercase : List[str]=3, __lowercase : Tuple=18, __lowercase : List[Any]=30, __lowercase : Tuple=400, __lowercase : Any=None, __lowercase : Optional[int]=True, __lowercase : List[str]=True, __lowercase : Union[str, Any]=None, ): lowercase__ = size if size is not None else {"height": 20, "width": 20} lowercase__ = parent lowercase__ = batch_size lowercase__ = num_channels lowercase__ = image_size lowercase__ = min_resolution lowercase__ = max_resolution lowercase__ = size lowercase__ = do_normalize lowercase__ = do_convert_rgb lowercase__ = [512, 1024, 2048, 4096] lowercase__ = patch_size if patch_size is not None else {"height": 16, "width": 16} def A__ ( self : List[str] ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def A__ ( self : Any ): lowercase__ = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg" lowercase__ = Image.open(requests.get(__lowercase, stream=__lowercase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Any =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Any ): lowercase__ = PixaStructImageProcessingTester(self ) @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Optional[Any] ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Optional[int] ): lowercase__ = self.image_processor_tester.prepare_dummy_image() lowercase__ = self.image_processing_class(**self.image_processor_dict ) lowercase__ = 2048 lowercase__ = image_processor(__lowercase, return_tensors="pt", max_patches=__lowercase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean(), torch.tensor(0.0606 ), atol=1e-3, rtol=1e-3 ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : int ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 lowercase__ = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__lowercase ): lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches lowercase__ = "Hello" lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase, header_text=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Tuple ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, numpify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, np.ndarray ) lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) def A__ ( self : Any ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase, torchify=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, torch.Tensor ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Optional[int] =PixaStructImageProcessor if is_vision_available() else None def A__ ( self : Optional[int] ): lowercase__ = PixaStructImageProcessingTester(self, num_channels=4 ) lowercase__ = 3 @property def A__ ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self : Dict ): lowercase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowercase, "do_normalize" ) ) self.assertTrue(hasattr(__lowercase, "do_convert_rgb" ) ) def A__ ( self : Union[str, Any] ): # Initialize image_processor lowercase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ = prepare_image_inputs(self.image_processor_tester, equal_resolution=__lowercase ) for image in image_inputs: self.assertIsInstance(__lowercase, Image.Image ) # Test not batched input lowercase__ = ( (self.image_processor_tester.patch_size["height"] * self.image_processor_tester.patch_size["width"]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase__ = image_processor( image_inputs[0], return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (1, max_patch, expected_hidden_dim), ) # Test batched lowercase__ = image_processor( __lowercase, return_tensors="pt", max_patches=__lowercase ).flattened_patches self.assertEqual( encoded_images.shape, (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim), )
37
0
'''simple docstring''' import os from math import logaa def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ = "base_exp.txt" ): lowercase__ = 0 lowercase__ = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) ) ): lowercase__ = list(map(SCREAMING_SNAKE_CASE_ , line.split("," ) ) ) if x * logaa(SCREAMING_SNAKE_CASE_ ) > largest: lowercase__ = x * logaa(SCREAMING_SNAKE_CASE_ ) lowercase__ = i + 1 return result if __name__ == "__main__": print(solution())
720
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = len(SCREAMING_SNAKE_CASE_ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ = 0 count += depth_first_search(SCREAMING_SNAKE_CASE_ , row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , row - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col - 1 , SCREAMING_SNAKE_CASE_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
37
0
from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : bool = False ): if radian_mode: return [magnitude * cos(SCREAMING_SNAKE_CASE_ ), magnitude * sin(SCREAMING_SNAKE_CASE_ )] return [magnitude * cos(radians(SCREAMING_SNAKE_CASE_ ) ), magnitude * sin(radians(SCREAMING_SNAKE_CASE_ ) )] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ : NDArray[floataa] , SCREAMING_SNAKE_CASE_ : NDArray[floataa] , SCREAMING_SNAKE_CASE_ : float = 10**-1 ): lowercase__ = cross(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = sum(SCREAMING_SNAKE_CASE_ ) return abs(SCREAMING_SNAKE_CASE_ ) < eps if __name__ == "__main__": # Test to check if it works lowercase_ = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) lowercase_ = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg lowercase_ = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) lowercase_ = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg lowercase_ = array([[0, -2000], [0, -1200], [0, 1_5600], [0, -1_2400]]) lowercase_ = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
721
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 for ch in input_str: lowercase__ = ord(SCREAMING_SNAKE_CASE_ ) lowercase__ = pow(2 , SCREAMING_SNAKE_CASE_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
37
0
'''simple docstring''' def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(_lowerCAmelCase , n - 1 , _lowerCAmelCase ) * a) % mod else: lowercase__ = binary_exponentiation(_lowerCAmelCase , n / 2 , _lowerCAmelCase ) return (b * b) % mod # a prime number lowercase_ = 701 lowercase_ = 10_0000_0000 lowercase_ = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
700
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = len(SCREAMING_SNAKE_CASE_ ) for i in range(SCREAMING_SNAKE_CASE_ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE_ ): if numbers[j] < numbers[i]: lowercase__ , lowercase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() lowercase_ = [int(item) for item in user_input.split(""",""")] print(exchange_sort(unsorted))
37
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = 0 lowercase__ = len(lowerCamelCase_ ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None lowercase__ = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCamelCase_ ): return None lowercase__ = sorted_collection[point] if current_item == item: return point else: if point < left: lowercase__ = left lowercase__ = point elif point > right: lowercase__ = right lowercase__ = point else: if item < current_item: lowercase__ = point - 1 else: lowercase__ = point + 1 return None def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None lowercase__ = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCamelCase_ ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) elif point > right: return interpolation_search_by_recursion(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , point - 1 ) else: return interpolation_search_by_recursion( lowerCamelCase_ , lowerCamelCase_ , point + 1 , lowerCamelCase_ ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if collection != sorted(lowerCamelCase_ ): raise ValueError("Collection must be ascending sorted" ) return True if __name__ == "__main__": import sys lowercase_ = 0 if debug == 1: lowercase_ = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("""Sequence must be ascending sorted to apply interpolation search""") lowercase_ = 67 lowercase_ = interpolation_search(collection, target) if result is not None: print(F'{target} found at positions: {result}') else: print("""Not found""")
701
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) lowercase__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 lowercase__ = 1 if upper_limit > 0: lowercase__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print("""\n********* Catalan Numbers Using Dynamic Programming ************\n""") print("""\n*** Enter -1 at any time to quit ***""") print("""\nEnter the upper limit (≥ 0) for the Catalan number sequence: """, end="""""") try: while True: lowercase_ = int(input().strip()) if N < 0: print("""\n********* Goodbye!! ************""") break else: print(F'The Catalan numbers from 0 through {N} are:') print(catalan_numbers(N)) print("""Try another upper limit for the sequence: """, end="""""") except (NameError, ValueError): print("""\n********* Invalid input, goodbye! ************\n""") import doctest doctest.testmod()
37
0
import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class _snake_case ( UpperCAmelCase_ , unittest.TestCase): UpperCamelCase__ : List[str] =BertTokenizer UpperCamelCase__ : Dict =BertTokenizerFast UpperCamelCase__ : Any =True UpperCamelCase__ : Any =True UpperCamelCase__ : int =filter_non_english def A__ ( self : Optional[Any] ): super().setUp() lowercase__ = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : int, __lowercase : Any ): lowercase__ = "UNwant\u00E9d,running" lowercase__ = "unwanted, running" return input_text, output_text def A__ ( self : int ): lowercase__ = self.tokenizer_class(self.vocab_file ) lowercase__ = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(_lowercase, ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ), [9, 6, 7, 12, 10, 11] ) def A__ ( self : Union[str, Any] ): if not self.test_rust_tokenizer: return lowercase__ = self.get_tokenizer() lowercase__ = self.get_rust_tokenizer() lowercase__ = "UNwant\u00E9d,running" lowercase__ = tokenizer.tokenize(_lowercase ) lowercase__ = rust_tokenizer.tokenize(_lowercase ) self.assertListEqual(_lowercase, _lowercase ) lowercase__ = tokenizer.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = rust_tokenizer.encode(_lowercase, add_special_tokens=_lowercase ) self.assertListEqual(_lowercase, _lowercase ) lowercase__ = self.get_rust_tokenizer() lowercase__ = tokenizer.encode(_lowercase ) lowercase__ = rust_tokenizer.encode(_lowercase ) self.assertListEqual(_lowercase, _lowercase ) # With lower casing lowercase__ = self.get_tokenizer(do_lower_case=_lowercase ) lowercase__ = self.get_rust_tokenizer(do_lower_case=_lowercase ) lowercase__ = "UNwant\u00E9d,running" lowercase__ = tokenizer.tokenize(_lowercase ) lowercase__ = rust_tokenizer.tokenize(_lowercase ) self.assertListEqual(_lowercase, _lowercase ) lowercase__ = tokenizer.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = rust_tokenizer.encode(_lowercase, add_special_tokens=_lowercase ) self.assertListEqual(_lowercase, _lowercase ) lowercase__ = self.get_rust_tokenizer() lowercase__ = tokenizer.encode(_lowercase ) lowercase__ = rust_tokenizer.encode(_lowercase ) self.assertListEqual(_lowercase, _lowercase ) def A__ ( self : List[Any] ): lowercase__ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ), ["ah", "\u535A", "\u63A8", "zz"] ) def A__ ( self : int ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ), ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ), ["hello"] ) def A__ ( self : List[Any] ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase, strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ), ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ), ["h\u00E9llo"] ) def A__ ( self : Any ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase, strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ), ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ), ["hello"] ) def A__ ( self : Tuple ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ), ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ), ["hello"] ) def A__ ( self : int ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : Tuple ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase, strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ), ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : Any ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase, strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ), ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : List[Any] ): lowercase__ = BasicTokenizer(do_lower_case=_lowercase, never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ), ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def A__ ( self : List[Any] ): lowercase__ = BasicTokenizer() lowercase__ = "a\n\'ll !!to?\'d of, can\'t." lowercase__ = ["a", "\'", "ll", "!", "!", "to", "?", "\'", "d", "of", ",", "can", "\'", "t", "."] self.assertListEqual(tokenizer.tokenize(_lowercase ), _lowercase ) def A__ ( self : List[Any] ): lowercase__ = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] lowercase__ = {} for i, token in enumerate(_lowercase ): lowercase__ = i lowercase__ = WordpieceTokenizer(vocab=_lowercase, unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ), [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ), ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ), ["[UNK]", "runn", "##ing"] ) def A__ ( self : int ): self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def A__ ( self : Optional[int] ): self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def A__ ( self : Optional[int] ): self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def A__ ( self : Union[str, Any] ): lowercase__ = self.get_tokenizer() lowercase__ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowercase ) for t in ["Test", "\xad", "test"]], [["[UNK]"], [], ["[UNK]"]] ) self.assertListEqual( [rust_tokenizer.tokenize(_lowercase ) for t in ["Test", "\xad", "test"]], [["[UNK]"], [], ["[UNK]"]] ) @slow def A__ ( self : List[str] ): lowercase__ = self.tokenizer_class.from_pretrained("bert-base-uncased" ) lowercase__ = tokenizer.encode("sequence builders", add_special_tokens=_lowercase ) lowercase__ = tokenizer.encode("multi-sequence build", add_special_tokens=_lowercase ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowercase ) lowercase__ = tokenizer.build_inputs_with_special_tokens(_lowercase, _lowercase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def A__ ( self : Any ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase, **_lowercase ) lowercase__ = F'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' lowercase__ = tokenizer_r.encode_plus( _lowercase, return_attention_mask=_lowercase, return_token_type_ids=_lowercase, return_offsets_mapping=_lowercase, add_special_tokens=_lowercase, ) lowercase__ = tokenizer_r.do_lower_case if hasattr(_lowercase, "do_lower_case" ) else False lowercase__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results], tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results], tokens["offset_mapping"] ) def A__ ( self : List[str] ): lowercase__ = ["的", "人", "有"] lowercase__ = "".join(_lowercase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase__ = True lowercase__ = self.tokenizer_class.from_pretrained(_lowercase, **_lowercase ) lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase, **_lowercase ) lowercase__ = tokenizer_p.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.convert_ids_to_tokens(_lowercase ) lowercase__ = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowercase, _lowercase ) self.assertListEqual(_lowercase, _lowercase ) lowercase__ = False lowercase__ = self.rust_tokenizer_class.from_pretrained(_lowercase, **_lowercase ) lowercase__ = self.tokenizer_class.from_pretrained(_lowercase, **_lowercase ) lowercase__ = tokenizer_r.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = tokenizer_p.encode(_lowercase, add_special_tokens=_lowercase ) lowercase__ = tokenizer_r.convert_ids_to_tokens(_lowercase ) lowercase__ = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that only the first Chinese character is not preceded by "##". lowercase__ = [ F'''##{token}''' if idx != 0 else token for idx, token in enumerate(_lowercase ) ] self.assertListEqual(_lowercase, _lowercase ) self.assertListEqual(_lowercase, _lowercase )
702
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase_ = { """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaPreLayerNormConfig""", """RobertaPreLayerNormOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaPreLayerNormForCausalLM""", """RobertaPreLayerNormForMaskedLM""", """RobertaPreLayerNormForMultipleChoice""", """RobertaPreLayerNormForQuestionAnswering""", """RobertaPreLayerNormForSequenceClassification""", """RobertaPreLayerNormForTokenClassification""", """RobertaPreLayerNormModel""", """RobertaPreLayerNormPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaPreLayerNormForCausalLM""", """TFRobertaPreLayerNormForMaskedLM""", """TFRobertaPreLayerNormForMultipleChoice""", """TFRobertaPreLayerNormForQuestionAnswering""", """TFRobertaPreLayerNormForSequenceClassification""", """TFRobertaPreLayerNormForTokenClassification""", """TFRobertaPreLayerNormMainLayer""", """TFRobertaPreLayerNormModel""", """TFRobertaPreLayerNormPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """FlaxRobertaPreLayerNormForCausalLM""", """FlaxRobertaPreLayerNormForMaskedLM""", """FlaxRobertaPreLayerNormForMultipleChoice""", """FlaxRobertaPreLayerNormForQuestionAnswering""", """FlaxRobertaPreLayerNormForSequenceClassification""", """FlaxRobertaPreLayerNormForTokenClassification""", """FlaxRobertaPreLayerNormModel""", """FlaxRobertaPreLayerNormPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
from scipy.stats import spearmanr import datasets lowercase_ = """\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n""" lowercase_ = """\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {'spearmanr': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results['spearmanr'])\n -0.7\n >>> print(round(results['spearmanr_pvalue'], 2))\n 0.19\n""" lowercase_ = r"""\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _snake_case ( datasets.Metric): def A__ ( self : Any ): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ), reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"], ) def A__ ( self : str, __lowercase : Dict, __lowercase : int, __lowercase : List[str]=False ): lowercase__ = spearmanr(lowercase_, lowercase_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
703
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") lowercase__ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowercase__ = model.state_dict() def to_tf_var_name(SCREAMING_SNAKE_CASE_ ): for patt, repl in iter(SCREAMING_SNAKE_CASE_ ): lowercase__ = name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return f'''bert/{name}''' def create_tf_var(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=SCREAMING_SNAKE_CASE_ , shape=tensor.shape , name=SCREAMING_SNAKE_CASE_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(SCREAMING_SNAKE_CASE_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(SCREAMING_SNAKE_CASE_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=SCREAMING_SNAKE_CASE_ , name=SCREAMING_SNAKE_CASE_ , session=SCREAMING_SNAKE_CASE_ ) tf.keras.backend.set_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase__ = session.run(SCREAMING_SNAKE_CASE_ ) print(f'''Successfully created {tf_name}: {np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(SCREAMING_SNAKE_CASE_ , os.path.join(SCREAMING_SNAKE_CASE_ , model_name.replace("-" , "_" ) + ".ckpt" ) ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_=None ): lowercase__ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help="Directory in which to save tensorflow model" ) lowercase__ = parser.parse_args(SCREAMING_SNAKE_CASE_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=SCREAMING_SNAKE_CASE_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
37
0
from pathlib import Path import fire def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = Path(SCREAMING_SNAKE_CASE_ ) lowercase__ = Path(SCREAMING_SNAKE_CASE_ ) dest_dir.mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) for path in src_dir.iterdir(): lowercase__ = [x.rstrip() for x in list(path.open().readlines() )][:n] lowercase__ = dest_dir.joinpath(path.name ) print(SCREAMING_SNAKE_CASE_ ) dest_path.open("w" ).write("\n".join(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": fire.Fire(minify)
704
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { """configuration_timesformer""": ["""TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimesformerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ """TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimesformerModel""", """TimesformerForVideoClassification""", """TimesformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
37
0
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowercase_ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowercase_ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ): lowercase__ = text.split(_lowerCamelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(_lowerCamelCase ) , _lowerCamelCase )] def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(_lowerCamelCase ): titles.append(title if title is not None else "" ) texts.append(_lowerCamelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = ctx_tokenizer( documents["title"] , documents["text"] , truncation=_lowerCamelCase , padding="longest" , return_tensors="pt" )["input_ids"] lowercase__ = ctx_encoder(input_ids.to(device=_lowerCamelCase ) , return_dict=_lowerCamelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ): logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowercase__ = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowercase__ = dataset.map(_lowerCamelCase , batched=_lowerCamelCase , num_proc=processing_args.num_proc ) # And compute the embeddings lowercase__ = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=_lowerCamelCase ) lowercase__ = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowercase__ = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space lowercase__ = dataset.map( partial(_lowerCamelCase , ctx_encoder=_lowerCamelCase , ctx_tokenizer=_lowerCamelCase ) , batched=_lowerCamelCase , batch_size=processing_args.batch_size , features=_lowerCamelCase , ) # And finally save your dataset lowercase__ = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(_lowerCamelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowercase__ = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=_lowerCamelCase ) # And save the index lowercase__ = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(_lowerCamelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class _snake_case : UpperCamelCase__ : Tuple =field( default=str(Path(__lowercase).parent / """test_run""" / """dummy-kb""" / """my_knowledge_dataset.csv""") , metadata={"""help""": """Path to a tab-separated csv file with columns \'title\' and \'text\'"""} , ) UpperCamelCase__ : int =field( default=__lowercase , metadata={"""help""": """Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'."""} , ) UpperCamelCase__ : Tuple =field( default="""facebook/rag-sequence-nq""" , metadata={"""help""": """The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\'"""} , ) UpperCamelCase__ : Tuple =field( default="""facebook/dpr-ctx_encoder-multiset-base""" , metadata={ """help""": ( """The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or""" """ \'facebook/dpr-ctx_encoder-multiset-base\'""" ) } , ) UpperCamelCase__ : Any =field( default=str(Path(__lowercase).parent / """test_run""" / """dummy-kb""") , metadata={"""help""": """Path to a directory where the dataset passages and the index will be saved"""} , ) @dataclass class _snake_case : UpperCamelCase__ : str =field( default=__lowercase , metadata={ """help""": """The number of processes to use to split the documents into passages. Default is single process.""" } , ) UpperCamelCase__ : Optional[Any] =field( default=1_6 , metadata={ """help""": """The batch size to use when computing the passages embeddings using the DPR context encoder.""" } , ) @dataclass class _snake_case : UpperCamelCase__ : Tuple =field( default=7_6_8 , metadata={"""help""": """The dimension of the embeddings to pass to the HNSW Faiss index."""} , ) UpperCamelCase__ : str =field( default=1_2_8 , metadata={ """help""": ( """The number of bi-directional links created for every new element during the HNSW index construction.""" ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowercase_ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowercase_ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowercase_ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
705
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() lowercase_ = { """bart""": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), """bert""": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-uncased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-large-cased-whole-word-masking-finetuned-squad""": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """bert-base-cased-finetuned-mrpc""": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """dpr""": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), """gpt2""": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlnet""": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm""": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """xlm-roberta""": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """transfo-xl""": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """openai-gpt""": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """roberta""": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """layoutlm""": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), """roberta-large-mnli""": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """camembert""": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """flaubert""": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert""": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """distilbert-base-distilled-squad""": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert""": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """lxmert-visual-feature-encoder""": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """ctrl""": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """albert""": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """t5""": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """electra""": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), """wav2vec2""": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=True ): if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) lowercase__ = config_class.from_json_file(SCREAMING_SNAKE_CASE_ ) lowercase__ = True lowercase__ = True print(f'''Building TensorFlow model from configuration: {config}''' ) lowercase__ = model_class(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): lowercase__ = cached_file( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: lowercase__ = load_pytorch_checkpoint_in_tfa_model(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if compare_with_pt_model: lowercase__ = tf_model(tf_model.dummy_inputs , training=SCREAMING_SNAKE_CASE_ ) # build the network lowercase__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" ) lowercase__ = pt_model_class.from_pretrained( pretrained_model_name_or_path=SCREAMING_SNAKE_CASE_ , config=SCREAMING_SNAKE_CASE_ , state_dict=SCREAMING_SNAKE_CASE_ ) with torch.no_grad(): lowercase__ = pt_model(**pt_model.dummy_inputs ) lowercase__ = pto[0].numpy() lowercase__ = tfo[0].numpy() lowercase__ = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(SCREAMING_SNAKE_CASE_ , save_format="h5" ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ): if args_model_type is None: lowercase__ = list(MODEL_CLASSES.keys() ) else: lowercase__ = [args_model_type] for j, model_type in enumerate(SCREAMING_SNAKE_CASE_ , start=1 ): print("=" * 100 ) print(f''' Converting model type {j}/{len(SCREAMING_SNAKE_CASE_ )}: {model_type}''' ) print("=" * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: lowercase__ = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: lowercase__ = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , start=1 ): print("-" * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue lowercase__ = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(SCREAMING_SNAKE_CASE_ )}: {model_shortcut_name} - model_type {model_type}''' ) print("-" * 100 ) if config_shortcut_name in aws_config_map: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = config_shortcut_name if model_shortcut_name in aws_model_maps: lowercase__ = cached_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , force_download=not use_cached_models ) else: lowercase__ = model_shortcut_name if os.path.isfile(SCREAMING_SNAKE_CASE_ ): lowercase__ = "converted_model" convert_pt_checkpoint_to_tf( model_type=SCREAMING_SNAKE_CASE_ , pytorch_checkpoint_path=SCREAMING_SNAKE_CASE_ , config_file=SCREAMING_SNAKE_CASE_ , tf_dump_path=os.path.join(SCREAMING_SNAKE_CASE_ , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=SCREAMING_SNAKE_CASE_ , ) if remove_cached_files: os.remove(SCREAMING_SNAKE_CASE_ ) os.remove(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_dump_path""", default=None, type=str, required=True, help="""Path to the output Tensorflow dump file.""" ) parser.add_argument( """--model_type""", default=None, type=str, help=( F'Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ' """convert all the models from AWS.""" ), ) parser.add_argument( """--pytorch_checkpoint_path""", default=None, type=str, help=( """Path to the PyTorch checkpoint path or shortcut name to download from AWS. """ """If not given, will download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--config_file""", default=None, type=str, help=( """The config json file corresponding to the pre-trained model. \n""" """This specifies the model architecture. If not given and """ """--pytorch_checkpoint_path is not given or is a shortcut name """ """use the configuration associated to the shortcut name on the AWS""" ), ) parser.add_argument( """--compare_with_pt_model""", action="""store_true""", help="""Compare Tensorflow and PyTorch model predictions.""" ) parser.add_argument( """--use_cached_models""", action="""store_true""", help="""Use cached models if possible instead of updating to latest checkpoint versions.""", ) parser.add_argument( """--remove_cached_files""", action="""store_true""", help="""Remove pytorch models after conversion (save memory when converting in batches).""", ) parser.add_argument("""--only_convert_finetuned_models""", action="""store_true""", help="""Only convert finetuned models.""") lowercase_ = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
37
0
lowercase_ = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowercase_ = concatenate_datasets lowercase_ = DownloadConfig lowercase_ = DownloadManager lowercase_ = DownloadMode lowercase_ = DownloadConfig lowercase_ = DownloadMode lowercase_ = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
706
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase_ = logging.get_logger(__name__) class _snake_case ( A_): UpperCamelCase__ : Union[str, Any] =['''pixel_values'''] def __init__( self : Tuple, __lowercase : Tuple = True, __lowercase : Tuple = None, __lowercase : Any = PIL.Image.BICUBIC, __lowercase : int = True, __lowercase : Any = None, __lowercase : Any = 1 / 255, __lowercase : int = True, __lowercase : Dict = True, __lowercase : str = None, __lowercase : Optional[Any] = None, **__lowercase : str, ): super().__init__(**__lowercase ) lowercase__ = size if size is not None else {"""height""": 256, """width""": 256} lowercase__ = get_size_dict(__lowercase ) lowercase__ = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} lowercase__ = get_size_dict(__lowercase, param_name="crop_size" ) lowercase__ = do_resize lowercase__ = size lowercase__ = resample lowercase__ = do_center_crop lowercase__ = crop_size lowercase__ = do_rescale lowercase__ = rescale_factor lowercase__ = do_normalize lowercase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowercase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self : List[str], __lowercase : str, __lowercase : List[Any], __lowercase : Optional[Any] = PIL.Image.BICUBIC, __lowercase : Optional[int] = None, **__lowercase : Any, ): lowercase__ = get_size_dict(__lowercase ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return resize( __lowercase, size=(size["height"], size["width"]), resample=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : Union[str, Any], __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any] = None, **__lowercase : List[str], ): lowercase__ = get_size_dict(__lowercase ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return center_crop(__lowercase, size=(size["height"], size["width"]), data_format=__lowercase, **__lowercase ) def A__ ( self : Any, __lowercase : Any, __lowercase : str, __lowercase : Optional[Any] = None, **__lowercase : Optional[int], ): return rescale(__lowercase, scale=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : Tuple, __lowercase : Union[str, Any], __lowercase : List[str], __lowercase : int, __lowercase : Optional[int] = None, **__lowercase : Tuple, ): return normalize(__lowercase, mean=__lowercase, std=__lowercase, data_format=__lowercase, **__lowercase ) def A__ ( self : List[Any], __lowercase : int, __lowercase : List[str] = None, __lowercase : Optional[int] = None, __lowercase : Optional[int]=None, __lowercase : Optional[int] = None, __lowercase : List[str] = None, __lowercase : List[Any] = None, __lowercase : List[Any] = None, __lowercase : List[str] = None, __lowercase : Any = None, __lowercase : Optional[int] = None, __lowercase : List[str] = None, __lowercase : List[str] = ChannelDimension.FIRST, **__lowercase : List[str], ): lowercase__ = do_resize if do_resize is not None else self.do_resize lowercase__ = resample if resample is not None else self.resample lowercase__ = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase__ = do_rescale if do_rescale is not None else self.do_rescale lowercase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ = do_normalize if do_normalize is not None else self.do_normalize lowercase__ = image_mean if image_mean is not None else self.image_mean lowercase__ = image_std if image_std is not None else self.image_std lowercase__ = size if size is not None else self.size lowercase__ = get_size_dict(__lowercase ) lowercase__ = crop_size if crop_size is not None else self.crop_size lowercase__ = get_size_dict(__lowercase, param_name="crop_size" ) lowercase__ = make_list_of_images(__lowercase ) if not valid_images(__lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. lowercase__ = [to_numpy_array(__lowercase ) for image in images] if do_resize: lowercase__ = [self.resize(image=__lowercase, size=__lowercase, resample=__lowercase ) for image in images] if do_center_crop: lowercase__ = [self.center_crop(image=__lowercase, size=__lowercase ) for image in images] if do_rescale: lowercase__ = [self.rescale(image=__lowercase, scale=__lowercase ) for image in images] if do_normalize: lowercase__ = [self.normalize(image=__lowercase, mean=__lowercase, std=__lowercase ) for image in images] lowercase__ = [to_channel_dimension_format(__lowercase, __lowercase ) for image in images] lowercase__ = {"""pixel_values""": images} return BatchFeature(data=__lowercase, tensor_type=__lowercase )
707
import pickle import numpy as np from matplotlib import pyplot as plt class _snake_case : def __init__( self : Tuple, __lowercase : Union[str, Any], __lowercase : int, __lowercase : Union[str, Any], __lowercase : str, __lowercase : List[Any], __lowercase : List[str]=0.2, __lowercase : List[str]=0.2 ): lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = bp_numa lowercase__ = conva_get[:2] lowercase__ = conva_get[2] lowercase__ = size_pa lowercase__ = rate_w lowercase__ = rate_t lowercase__ = [ np.mat(-1 * np.random.rand(self.conva[0], self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = np.mat(-1 * np.random.rand(self.num_bpa, self.num_bpa ) + 0.5 ) lowercase__ = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self : Any, __lowercase : List[str] ): # save model dict with pickle lowercase__ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(__lowercase, "wb" ) as f: pickle.dump(__lowercase, __lowercase ) print(F'''Model saved: {save_path}''' ) @classmethod def A__ ( cls : Dict, __lowercase : Union[str, Any] ): # read saved model with open(__lowercase, "rb" ) as f: lowercase__ = pickle.load(__lowercase ) # noqa: S301 lowercase__ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) lowercase__ = model_dic.get("size_pooling1" ) lowercase__ = model_dic.get("num_bp1" ) lowercase__ = model_dic.get("num_bp2" ) lowercase__ = model_dic.get("num_bp3" ) lowercase__ = model_dic.get("rate_weight" ) lowercase__ = model_dic.get("rate_thre" ) # create model instance lowercase__ = CNN(__lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase, __lowercase ) # modify model parameter lowercase__ = model_dic.get("w_conv1" ) lowercase__ = model_dic.get("wkj" ) lowercase__ = model_dic.get("vji" ) lowercase__ = model_dic.get("thre_conv1" ) lowercase__ = model_dic.get("thre_bp2" ) lowercase__ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self : str, __lowercase : List[Any] ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self : List[str], __lowercase : Optional[Any] ): return round(__lowercase, 3 ) def A__ ( self : Optional[Any], __lowercase : Dict, __lowercase : Optional[int], __lowercase : Optional[int], __lowercase : Optional[Any], __lowercase : str ): # convolution process lowercase__ = convs[0] lowercase__ = convs[1] lowercase__ = np.shape(__lowercase )[0] # get the data slice of original image data, data_focus lowercase__ = [] for i_focus in range(0, size_data - size_conv + 1, __lowercase ): for j_focus in range(0, size_data - size_conv + 1, __lowercase ): lowercase__ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(__lowercase ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ = [] lowercase__ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(__lowercase ): lowercase__ = [] for i_focus in range(len(__lowercase ) ): lowercase__ = ( np.sum(np.multiply(data_focus[i_focus], w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape( __lowercase, __lowercase ) data_featuremap.append(__lowercase ) # expanding the data slice to One dimenssion lowercase__ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(__lowercase ) ) lowercase__ = np.asarray(__lowercase ) return focus_list, data_featuremap def A__ ( self : List[Any], __lowercase : Any, __lowercase : List[Any], __lowercase : Union[str, Any]="average_pool" ): # pooling process lowercase__ = len(featuremaps[0] ) lowercase__ = int(size_map / size_pooling ) lowercase__ = [] for i_map in range(len(__lowercase ) ): lowercase__ = featuremaps[i_map] lowercase__ = [] for i_focus in range(0, __lowercase, __lowercase ): for j_focus in range(0, __lowercase, __lowercase ): lowercase__ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(__lowercase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(__lowercase ) ) lowercase__ = np.asmatrix(__lowercase ).reshape(__lowercase, __lowercase ) featuremap_pooled.append(__lowercase ) return featuremap_pooled def A__ ( self : str, __lowercase : Optional[Any] ): # expanding three dimension data to one dimension list lowercase__ = [] for i in range(len(__lowercase ) ): lowercase__ = np.shape(data[i] ) lowercase__ = data[i].reshape(1, shapes[0] * shapes[1] ) lowercase__ = data_listed.getA().tolist()[0] data_expanded.extend(__lowercase ) lowercase__ = np.asarray(__lowercase ) return data_expanded def A__ ( self : Optional[int], __lowercase : Optional[int] ): # expanding matrix to one dimension list lowercase__ = np.asarray(__lowercase ) lowercase__ = np.shape(__lowercase ) lowercase__ = data_mat.reshape(1, shapes[0] * shapes[1] ) return data_expanded def A__ ( self : str, __lowercase : Tuple, __lowercase : List[Any], __lowercase : Any, __lowercase : Union[str, Any], __lowercase : Tuple ): lowercase__ = [] lowercase__ = 0 for i_map in range(__lowercase ): lowercase__ = np.ones((size_map, size_map) ) for i in range(0, __lowercase, __lowercase ): for j in range(0, __lowercase, __lowercase ): lowercase__ = pd_pool[ i_pool ] lowercase__ = i_pool + 1 lowercase__ = np.multiply( __lowercase, np.multiply(out_map[i_map], (1 - out_map[i_map]) ) ) pd_all.append(__lowercase ) return pd_all def A__ ( self : Tuple, __lowercase : int, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : List[str]=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(__lowercase )) ) print((" - - Shape: Teach_Data ", np.shape(__lowercase )) ) lowercase__ = 0 lowercase__ = [] lowercase__ = 1_0000 while rp < n_repeat and mse >= error_accuracy: lowercase__ = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(__lowercase ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ = np.asmatrix(datas_train[p] ) lowercase__ = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = np.shape(__lowercase ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = np.dot(__lowercase, self.vji.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = np.dot(__lowercase, self.wkj.T ) - self.thre_bpa lowercase__ = self.sig(__lowercase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ = np.multiply( (data_teach - bp_outa), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.multiply( np.dot(__lowercase, self.wkj ), np.multiply(__lowercase, (1 - bp_outa) ) ) lowercase__ = np.dot(__lowercase, self.vji ) lowercase__ = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ = pd_conva_pooled.T.getA().tolist() lowercase__ = self._calculate_gradient_from_pool( __lowercase, __lowercase, shape_featuremapa[0], shape_featuremapa[1], self.size_poolinga, ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ = self.rate_weight * np.dot(__lowercase, __lowercase ) lowercase__ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ = rp + 1 lowercase__ = error_count / patterns all_mse.append(__lowercase ) def draw_error(): lowercase__ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(__lowercase, "+-" ) plt.plot(__lowercase, "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(__lowercase, alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self : List[str], __lowercase : Optional[int] ): # model predict lowercase__ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(__lowercase )) ) for p in range(len(__lowercase ) ): lowercase__ = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) lowercase__ = self._expand(__lowercase ) lowercase__ = data_bp_input lowercase__ = bp_outa * self.vji.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) lowercase__ = bp_outa * self.wkj.T - self.thre_bpa lowercase__ = self.sig(__lowercase ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ = [list(map(self.do_round, __lowercase ) ) for each in produce_out] return np.asarray(__lowercase ) def A__ ( self : int, __lowercase : Any ): # return the data of image after convoluting process so we can check it out lowercase__ = np.asmatrix(__lowercase ) lowercase__ , lowercase__ = self.convolute( __lowercase, self.conva, self.w_conva, self.thre_conva, conv_step=self.step_conva, ) lowercase__ = self.pooling(__lowercase, self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
37
0
import requests def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = {'Content-Type': 'application/json'} lowercase__ = requests.post(lowerCAmelCase__ , json={"text": message_body} , headers=lowerCAmelCase__ ) if response.status_code != 200: lowercase__ = ( 'Request to slack returned an error ' f'''{response.status_code}, the response is:\n{response.text}''' ) raise ValueError(lowerCAmelCase__ ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message("""<YOUR MESSAGE BODY>""", """<SLACK CHANNEL URL>""")
708
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = "huggingface/label-files" lowercase__ = "imagenet-1k-id2label.json" lowercase__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) lowercase__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowercase__ = {v: k for k, v in idalabel.items()} lowercase__ = "std_conv" if "bit" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowercase__ = BitConfig( conv_layer=SCREAMING_SNAKE_CASE_ , num_labels=1000 , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if "stem.conv" in name: lowercase__ = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: lowercase__ = name.replace("blocks" , "layers" ) if "head.fc" in name: lowercase__ = name.replace("head.fc" , "classifier.1" ) if name.startswith("norm" ): lowercase__ = "bit." + name if "bit" not in name and "classifier" not in name: lowercase__ = "bit.encoder." + name return name def __lowerCAmelCase ( ): lowercase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): lowercase__ = get_config(SCREAMING_SNAKE_CASE_ ) # load original model from timm lowercase__ = create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() # load state_dict of original model lowercase__ = timm_model.state_dict() for key in state_dict.copy().keys(): lowercase__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) lowercase__ = val.squeeze() if "head" in key else val # load HuggingFace model lowercase__ = BitForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # create image processor lowercase__ = create_transform(**resolve_data_config({} , model=SCREAMING_SNAKE_CASE_ ) ) lowercase__ = transform.transforms lowercase__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowercase__ = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE_ , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=SCREAMING_SNAKE_CASE_ , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=SCREAMING_SNAKE_CASE_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowercase__ = prepare_img() lowercase__ = transform(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ) lowercase__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # verify logits with torch.no_grad(): lowercase__ = model(SCREAMING_SNAKE_CASE_ ) lowercase__ = outputs.logits print("Logits:" , logits[0, :3] ) print("Predicted class:" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowercase__ = timm_model(SCREAMING_SNAKE_CASE_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE_ , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(f'''Saving model {model_name} and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(f'''Pushing model {model_name} and processor to the hub''' ) model.push_to_hub(f'''ybelkada/{model_name}''' ) processor.push_to_hub(f'''ybelkada/{model_name}''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""resnetv2_50x1_bitm""", type=str, help="""Name of the BiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub.""", ) lowercase_ = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
37
0
from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class _snake_case ( a__): UpperCamelCase__ : Tuple ="Salesforce/blip-image-captioning-base" UpperCamelCase__ : List[Any] =( "This is a tool that generates a description of an image. It takes an input named `image` which should be the " "image to caption, and returns a text that contains the description in English." ) UpperCamelCase__ : List[str] ="image_captioner" UpperCamelCase__ : Optional[int] =AutoModelForVisionaSeq UpperCamelCase__ : Union[str, Any] =["image"] UpperCamelCase__ : Optional[int] =["text"] def __init__( self : Union[str, Any], *__lowercase : Dict, **__lowercase : List[Any] ): requires_backends(self, ["vision"] ) super().__init__(*lowerCamelCase_, **lowerCamelCase_ ) def A__ ( self : Union[str, Any], __lowercase : Dict ): return self.pre_processor(images=lowerCamelCase_, return_tensors="pt" ) def A__ ( self : Any, __lowercase : Tuple ): return self.model.generate(**lowerCamelCase_ ) def A__ ( self : Union[str, Any], __lowercase : List[str] ): return self.pre_processor.batch_decode(lowerCamelCase_, skip_special_tokens=lowerCamelCase_ )[0].strip()
709
import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _snake_case ( lowercase__): def __init__( self : Optional[Any], __lowercase : str = "▁", __lowercase : bool = True, __lowercase : Union[str, AddedToken] = "<unk>", __lowercase : Union[str, AddedToken] = "</s>", __lowercase : Union[str, AddedToken] = "<pad>", ): lowercase__ = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } lowercase__ = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowercase__ = token_dict["token"] lowercase__ = Tokenizer(Unigram() ) lowercase__ = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ), " " ), normalizers.Lowercase(), ] ) lowercase__ = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) lowercase__ = decoders.Metaspace(replacement=__lowercase, add_prefix_space=__lowercase ) lowercase__ = TemplateProcessing( single=F'''$A {self.special_tokens["eos"]["token"]}''', special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])], ) lowercase__ = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase, __lowercase ) def A__ ( self : Union[str, Any], __lowercase : Union[str, List[str]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) if isinstance(__lowercase, __lowercase ): lowercase__ = [files] self._tokenizer.train(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : List[Any], __lowercase : Union[Iterator[str], Iterator[Iterator[str]]], __lowercase : int = 8000, __lowercase : bool = True, ): lowercase__ = trainers.UnigramTrainer( vocab_size=__lowercase, special_tokens=self.special_tokens_list, show_progress=__lowercase, ) self._tokenizer.train_from_iterator(__lowercase, trainer=__lowercase ) self.add_unk_id() def A__ ( self : str ): lowercase__ = json.loads(self._tokenizer.to_str() ) lowercase__ = self.special_tokens["unk"]["id"] lowercase__ = Tokenizer.from_str(json.dumps(__lowercase ) )
37
0
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values lowercase_ = argparse.ArgumentParser() parser.add_argument("""--user""", type=str, default="""ubuntu""") parser.add_argument("""--host""", type=str, default="""localhost""") parser.add_argument("""--key_path""", type=str, default=None) parser.add_argument("""--instance""", type=str, default="""V100:1""") parser.add_argument("""--provider""", type=str, default="""cheapest""") parser.add_argument("""--use_spot""", type=bool, default=False) parser.add_argument("""--example""", type=str, default="""pytorch/text-generation/run_generation.py""") lowercase_ , lowercase_ = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("""Cannot specify both BYO and on-demand cluster args""") lowercase_ = rh.cluster( name="""rh-cluster""", ips=[args.host], ssh_creds={"""ssh_user""": args.user, """ssh_private_key""": args.key_path} ) else: lowercase_ = rh.cluster( name="""rh-cluster""", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) lowercase_ = args.example.rsplit("""/""", 1)[0] # Set up remote environment cluster.install_packages(["""pip:./"""]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F'pip install -r transformers/examples/{example_dir}/requirements.txt']) cluster.run(["""pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"""]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F'python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
710
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowercase__ = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } lowercase__ = f'''{src_lang}-{tgt_lang}''' lowercase__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) lowercase__ = os.path.join(SCREAMING_SNAKE_CASE_ , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
37
0
def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE_ ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): lowercase__ = f'''Input value of [number={number}] must be an integer''' raise TypeError(__lowerCAmelCase ) if number < 0: return False lowercase__ = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
711
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =TransfoXLTokenizer UpperCamelCase__ : List[Any] =False UpperCamelCase__ : List[Any] =False def A__ ( self : Union[str, Any] ): super().setUp() lowercase__ = [ "<unk>", "[CLS]", "[SEP]", "want", "unwanted", "wa", "un", "running", ",", "low", "l", ] lowercase__ = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def A__ ( self : Union[str, Any], **__lowercase : Any ): lowercase__ = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname, **__lowercase ) def A__ ( self : Tuple, __lowercase : Optional[int] ): lowercase__ = "<unk> UNwanted , running" lowercase__ = "<unk> unwanted, running" return input_text, output_text def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(vocab_file=self.vocab_file, lower_case=__lowercase ) lowercase__ = tokenizer.tokenize("<unk> UNwanted , running" ) self.assertListEqual(__lowercase, ["<unk>", "unwanted", ",", "running"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase ), [0, 4, 8, 7] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["hello", "!", "how", "are", "you", "?"] ) def A__ ( self : Tuple ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo ! how \n Are yoU ? " ), ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def A__ ( self : str ): lowercase__ = TransfoXLTokenizer(lower_case=__lowercase ) lowercase__ = "Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?" lowercase__ = [ "Hello", "(", "bracket", ")", "and", "side", "@-@", "scrolled", "[", "and", "]", "Henry", "'s", "$", "5", "@,@", "000", "with", "3", "@.@", "34", "m", ".", "What", "'s", "up", "!", "?", ] self.assertListEqual(tokenizer.tokenize(__lowercase ), __lowercase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowercase ), __lowercase ) def A__ ( self : List[str] ): lowercase__ = self.get_tokenizer() lowercase__ = len(__lowercase ) tokenizer.add_tokens(["new1", "new2"] ) tokenizer.move_added_token("new1", 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowercase ), original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("new1" ), [1] ) self.assertEqual(tokenizer.decode([1] ), "new1" )
37
0
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": lowercase_ = """%20""".join(argv[1:]) if len(argv) > 1 else quote(str(input("""Search: """))) print("""Googling.....""") lowercase_ = F'https://www.google.com/search?q={query}&num=100' lowercase_ = requests.get( url, headers={"""User-Agent""": str(UserAgent().random)}, ) try: lowercase_ = ( BeautifulSoup(res.text, """html.parser""") .find("""div""", attrs={"""class""": """yuRUbf"""}) .find("""a""") .get("""href""") ) except AttributeError: lowercase_ = parse_qs( BeautifulSoup(res.text, """html.parser""") .find("""div""", attrs={"""class""": """kCrYT"""}) .find("""a""") .get("""href""") )["""url"""][0] webbrowser.open(link)
712
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def __lowerCAmelCase ( ): lowercase__ = HfArgumentParser(SCREAMING_SNAKE_CASE_ ) lowercase__ = parser.parse_args_into_dataclasses()[0] lowercase__ = TensorFlowBenchmark(args=SCREAMING_SNAKE_CASE_ ) try: lowercase__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowercase__ = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowercase__ = " ".join(str(SCREAMING_SNAKE_CASE_ ).split(" " )[:-1] ) lowercase__ = "" lowercase__ = eval(str(SCREAMING_SNAKE_CASE_ ).split(" " )[-1] ) lowercase__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowercase__ = full_error_msg + begin_error_msg + str(SCREAMING_SNAKE_CASE_ ) raise ValueError(SCREAMING_SNAKE_CASE_ ) benchmark.run() if __name__ == "__main__": main()
37
0