code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> list[list[int]]: UpperCamelCase : list[list[int]] = [] create_all_state(1 , _lowerCAmelCase , _lowerCAmelCase , [] , _lowerCAmelCase ) return result def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> None: if level == 0: total_list.append(current_list[:] ) return for i in range(_lowerCAmelCase , total_number - level + 2 ): current_list.append(_lowerCAmelCase ) create_all_state(i + 1 , _lowerCAmelCase , level - 1 , _lowerCAmelCase , _lowerCAmelCase ) current_list.pop() def A_ ( _lowerCAmelCase ) -> None: for i in total_list: print(*_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = 4 __lowerCamelCase : Any = 2 __lowerCamelCase : List[str] = generate_all_combinations(n, k) print_all_state(total_list)
38
from __future__ import annotations __lowerCamelCase : Optional[int] = """Muhammad Umer Farooq""" __lowerCamelCase : Tuple = """MIT""" __lowerCamelCase : Optional[int] = """1.0.0""" __lowerCamelCase : int = """Muhammad Umer Farooq""" __lowerCamelCase : Optional[int] = """contact@muhammadumerfarooq.me""" __lowerCamelCase : Dict = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : list[str] = [] UpperCamelCase : str = domain def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCamelCase : Any = parse.urljoin(self.domain , A_ ) self.urls.append(A_ ) def A_ ( _lowerCAmelCase ) -> str: return ".".join(get_sub_domain_name(_lowerCAmelCase ).split("." )[-2:] ) def A_ ( _lowerCAmelCase ) -> str: return parse.urlparse(_lowerCAmelCase ).netloc def A_ ( _lowerCAmelCase = "https://github.com" ) -> list[str]: UpperCamelCase : int = get_domain_name(_lowerCAmelCase ) # Initialize the parser UpperCamelCase : str = Parser(_lowerCAmelCase ) try: # Open URL UpperCamelCase : int = requests.get(_lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCamelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCamelCase : Optional[Any] = requests.get(_lowerCAmelCase ) # Get the valid email. UpperCamelCase : Optional[int] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = emails_from_url("""https://github.com""") print(f"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
38
1
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCamelCase : Dict = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class A__ ( unittest.TestCase ): def __UpperCamelCase( self , A_ , A_ = None , A_ = None , A_ = None , A_ = True , ): '''simple docstring''' UpperCamelCase : Union[str, Any] = [file for file in os.listdir(A_ ) if os.path.isfile(os.path.join(A_ , A_ ) )] if identifier is not None: UpperCamelCase : int = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(A_ , A_ ): for n_ in n_identifier: UpperCamelCase : Any = [file for file in files if n_ not in file] else: UpperCamelCase : Optional[int] = [file for file in files if n_identifier not in file] UpperCamelCase : Dict = ignore_files or [] ignore_files.append("__init__.py" ) UpperCamelCase : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , A_ ) if only_modules: UpperCamelCase : str = file.split("." )[0] try: UpperCamelCase : Union[str, Any] = getattr(A_ , A_ ) UpperCamelCase : int = doctest.DocTestSuite(A_ ) UpperCamelCase : Union[str, Any] = unittest.TextTestRunner().run(A_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCamelCase : Tuple = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = Path("src/transformers" ) UpperCamelCase : Dict = "modeling" UpperCamelCase : Union[str, Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(A_ , identifier=A_ , ignore_files=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = Path("src/transformers" ) UpperCamelCase : Tuple = "tokenization" self.analyze_directory(A_ , identifier=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = Path("src/transformers" ) UpperCamelCase : Optional[int] = "configuration" self.analyze_directory(A_ , identifier=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = Path("src/transformers" ) UpperCamelCase : Union[str, Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(A_ , n_identifier=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = Path("docs/source" ) UpperCamelCase : Optional[Any] = ["favicon.ico"] self.analyze_directory(A_ , ignore_files=A_ , only_modules=A_ )
38
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> list[int]: UpperCamelCase : Optional[Any] = [True] * limit UpperCamelCase : Optional[Any] = False UpperCamelCase : List[str] = False UpperCamelCase : Tuple = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCamelCase : Optional[Any] = i * 2 while index < limit: UpperCamelCase : int = False UpperCamelCase : Optional[int] = index + i UpperCamelCase : Any = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def A_ ( _lowerCAmelCase = 100_0000 ) -> int: UpperCamelCase : Union[str, Any] = prime_sieve(_lowerCAmelCase ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCamelCase : Dict = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCamelCase : int = j - i UpperCamelCase : Dict = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
38
1
import sys import turtle def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> tuple[float, float]: return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> None: my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(_lowerCAmelCase , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , depth - 1 ) triangle(_lowerCAmelCase , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , depth - 1 ) triangle(_lowerCAmelCase , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , get_mid(_lowerCAmelCase , _lowerCAmelCase ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( """Correct format for using this script: """ """python fractals.py <int:depth_for_fractal>""" ) __lowerCamelCase : int = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor("""red""") __lowerCamelCase : Tuple = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
38
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __snake_case ): def __init__( self , A_ , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , A_ = None , **A_ , ): '''simple docstring''' super().__init__( features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , num_proc=A_ , **A_ , ) UpperCamelCase : Optional[int] = Generator( cache_dir=A_ , features=A_ , generator=A_ , gen_kwargs=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' if self.streaming: UpperCamelCase : Optional[Any] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: UpperCamelCase : Union[str, Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , num_proc=self.num_proc , ) UpperCamelCase : int = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset
38
1
from PIL import Image def A_ ( _lowerCAmelCase ) -> Image: UpperCamelCase , UpperCamelCase : List[Any] = image.size UpperCamelCase : Union[str, Any] = 0 UpperCamelCase : List[str] = image.load() for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(_lowerCAmelCase ): for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
38
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : Tuple = """▁""" __lowerCamelCase : Optional[Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} __lowerCamelCase : List[str] = { """vocab_file""": { """facebook/mbart-large-en-ro""": ( """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model""" ), """facebook/mbart-large-cc25""": ( """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model""" ), } } __lowerCamelCase : Dict = { """facebook/mbart-large-en-ro""": 1024, """facebook/mbart-large-cc25""": 1024, } # fmt: off __lowerCamelCase : int = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN"""] class A__ ( __snake_case ): _UpperCAmelCase :Tuple = VOCAB_FILES_NAMES _UpperCAmelCase :int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[int] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :Any = ['input_ids', 'attention_mask'] _UpperCAmelCase :List[int] = [] _UpperCAmelCase :List[int] = [] def __init__( self , A_ , A_="<s>" , A_="</s>" , A_="</s>" , A_="<s>" , A_="<unk>" , A_="<pad>" , A_="<mask>" , A_=None , A_=None , A_=None , A_ = None , A_=None , **A_ , ): '''simple docstring''' UpperCamelCase : int = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else mask_token UpperCamelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A_ , eos_token=A_ , unk_token=A_ , sep_token=A_ , cls_token=A_ , pad_token=A_ , mask_token=A_ , tokenizer_file=A_ , src_lang=A_ , tgt_lang=A_ , additional_special_tokens=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) UpperCamelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(A_ ) ) UpperCamelCase : Optional[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCamelCase : Optional[int] = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCamelCase : Any = 1 UpperCamelCase : int = len(self.sp_model ) UpperCamelCase : str = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(A_ ) } UpperCamelCase : List[str] = {v: k for k, v in self.lang_code_to_id.items()} UpperCamelCase : Union[str, Any] = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCamelCase : Union[str, Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCamelCase : Any = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCamelCase : int = src_lang if src_lang is not None else "en_XX" UpperCamelCase : str = self.lang_code_to_id[self._src_lang] UpperCamelCase : Optional[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ): '''simple docstring''' UpperCamelCase : Dict = self.__dict__.copy() UpperCamelCase : Tuple = None UpperCamelCase : str = self.sp_model.serialized_model_proto() return state def __setstate__( self , A_ ): '''simple docstring''' UpperCamelCase : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCamelCase : Optional[int] = {} UpperCamelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __UpperCamelCase( self ): '''simple docstring''' return self._src_lang @src_lang.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __UpperCamelCase( self , A_ , A_ = None , A_ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) UpperCamelCase : int = [1] * len(self.prefix_tokens ) UpperCamelCase : List[Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(A_ )) + suffix_ones return prefix_ones + ([0] * len(A_ )) + ([0] * len(A_ )) + suffix_ones def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Dict = [self.sep_token_id] UpperCamelCase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __UpperCamelCase( self , A_ , A_ , A_ , A_ , **A_ ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCamelCase : Tuple = src_lang UpperCamelCase : Tuple = self(A_ , add_special_tokens=A_ , return_tensors=A_ , **A_ ) UpperCamelCase : Optional[int] = self.convert_tokens_to_ids(A_ ) UpperCamelCase : Dict = tgt_lang_id return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __UpperCamelCase( self , A_ ): '''simple docstring''' return self.sp_model.encode(A_ , out_type=A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCamelCase : Optional[Any] = self.sp_model.PieceToId(A_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __UpperCamelCase( self , A_ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = "".join(A_ ).replace(A_ , " " ).strip() return out_string def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' if not os.path.isdir(A_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase : List[str] = os.path.join( A_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , "wb" ) as fi: UpperCamelCase : str = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,) def __UpperCamelCase( self , A_ , A_ = "en_XX" , A_ = None , A_ = "ro_RO" , **A_ , ): '''simple docstring''' UpperCamelCase : str = src_lang UpperCamelCase : Dict = tgt_lang return super().prepare_seqaseq_batch(A_ , A_ , **A_ ) def __UpperCamelCase( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def __UpperCamelCase( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.lang_code_to_id[src_lang] UpperCamelCase : List[str] = [] UpperCamelCase : Optional[Any] = [self.eos_token_id, self.cur_lang_code] def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = self.lang_code_to_id[lang] UpperCamelCase : Any = [] UpperCamelCase : Dict = [self.eos_token_id, self.cur_lang_code]
38
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
1
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = LEDTokenizer _UpperCAmelCase :Union[str, Any] = LEDTokenizerFast _UpperCAmelCase :Tuple = True def __UpperCamelCase( self ): '''simple docstring''' super().setUp() UpperCamelCase : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] UpperCamelCase : Union[str, Any] = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase : Dict = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCamelCase : Any = {"unk_token": "<unk>"} UpperCamelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(A_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(A_ ) ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' return "lower newer", "lower newer" @cached_property def __UpperCamelCase( self ): '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def __UpperCamelCase( self ): '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] UpperCamelCase : Optional[Any] = [0, 250, 251, 1_7818, 13, 3_9186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : Optional[Any] = tokenizer(A_ , max_length=len(A_ ) , padding=A_ , return_tensors="pt" ) self.assertIsInstance(A_ , A_ ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) UpperCamelCase : List[Any] = batch.input_ids.tolist()[0] self.assertListEqual(A_ , A_ ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : Tuple = tokenizer(A_ , padding=A_ , return_tensors="pt" ) self.assertIn("input_ids" , A_ ) self.assertIn("attention_mask" , A_ ) self.assertNotIn("labels" , A_ ) self.assertNotIn("decoder_attention_mask" , A_ ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : Union[str, Any] = tokenizer(text_target=A_ , max_length=32 , padding="max_length" , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : List[str] = tokenizer( ["I am a small frog" * 1024, "I am a small frog"] , padding=A_ , truncation=A_ , return_tensors="pt" ) self.assertIsInstance(A_ , A_ ) self.assertEqual(batch.input_ids.shape , (2, 5122) ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = ["A long paragraph for summarization."] UpperCamelCase : List[str] = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : Optional[int] = tokenizer(A_ , return_tensors="pt" ) UpperCamelCase : Tuple = tokenizer(text_target=A_ , return_tensors="pt" ) UpperCamelCase : str = inputs["input_ids"] UpperCamelCase : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCamelCase : Any = ["Summary of the text.", "Another summary."] UpperCamelCase : int = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] UpperCamelCase : Dict = tokenizer(A_ , padding=A_ ) UpperCamelCase : Union[str, Any] = [[0] * len(A_ ) for x in encoded_output["input_ids"]] UpperCamelCase : Tuple = tokenizer.pad(A_ ) self.assertSequenceEqual(outputs["global_attention_mask"] , A_ ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCamelCase : Dict = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) UpperCamelCase : Tuple = self.tokenizer_class.from_pretrained(A_ , **A_ ) UpperCamelCase : Optional[Any] = "A, <mask> AllenNLP sentence." UpperCamelCase : Union[str, Any] = tokenizer_r.encode_plus(A_ , add_special_tokens=A_ , return_token_type_ids=A_ ) UpperCamelCase : Tuple = tokenizer_p.encode_plus(A_ , add_special_tokens=A_ , return_token_type_ids=A_ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) UpperCamelCase : str = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) UpperCamelCase : Optional[Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( A_ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( A_ , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
38
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
1
from __future__ import annotations import math def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> float: UpperCamelCase : Dict = u for i in range(1 , _lowerCAmelCase ): UpperCamelCase : Optional[Any] = temp * (u - i) return temp def A_ ( ) -> None: UpperCamelCase : Optional[Any] = int(input("enter the numbers of values: " ) ) UpperCamelCase : list[list[float]] = [] for _ in range(_lowerCAmelCase ): y.append([] ) for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): y[i].append(_lowerCAmelCase ) UpperCamelCase : str = 0 print("enter the values of parameters in a list: " ) UpperCamelCase : Tuple = list(map(_lowerCAmelCase , input().split() ) ) print("enter the values of corresponding parameters: " ) for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = float(input() ) UpperCamelCase : Optional[int] = int(input("enter the value to interpolate: " ) ) UpperCamelCase : int = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , _lowerCAmelCase ): for j in range(n - i ): UpperCamelCase : Optional[Any] = y[j + 1][i - 1] - y[j][i - 1] UpperCamelCase : int = y[0][0] for i in range(1 , _lowerCAmelCase ): summ += (ucal(_lowerCAmelCase , _lowerCAmelCase ) * y[0][i]) / math.factorial(_lowerCAmelCase ) print(F"""the value at {value} is {summ}""" ) if __name__ == "__main__": main()
38
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
1
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class A__ : def __init__( self , A_ , A_=2 , A_=32 , A_=16 , A_=3 , A_=True , A_=True , A_=32 , A_=4 , A_=[0, 1, 2, 3] , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=0.02 , A_=3 , A_=[1, 384, 24, 24] , A_=True , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : Tuple = image_size UpperCamelCase : Optional[int] = patch_size UpperCamelCase : str = num_channels UpperCamelCase : Tuple = is_training UpperCamelCase : Dict = use_labels UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : Optional[int] = num_hidden_layers UpperCamelCase : str = backbone_out_indices UpperCamelCase : Tuple = num_attention_heads UpperCamelCase : int = intermediate_size UpperCamelCase : str = hidden_act UpperCamelCase : List[Any] = hidden_dropout_prob UpperCamelCase : Any = attention_probs_dropout_prob UpperCamelCase : Dict = initializer_range UpperCamelCase : Dict = num_labels UpperCamelCase : Union[str, Any] = backbone_featmap_shape UpperCamelCase : Dict = scope UpperCamelCase : Any = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) UpperCamelCase : Any = (image_size // patch_size) ** 2 UpperCamelCase : Optional[int] = num_patches + 1 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : Optional[int] = None if self.use_labels: UpperCamelCase : str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase : Tuple = self.get_config() return config, pixel_values, labels def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [96, 192, 384, 768], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=A_ , backbone_featmap_shape=self.backbone_featmap_shape , ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = DPTModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : List[Any] = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = self.num_labels UpperCamelCase : Dict = DPTForDepthEstimation(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = self.num_labels UpperCamelCase : int = DPTForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Dict = model(A_ , labels=A_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase : Optional[Any] = config_and_inputs UpperCamelCase : List[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Tuple = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () _UpperCAmelCase :List[str] = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :str = False _UpperCAmelCase :Dict = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = DPTModelTester(self ) UpperCamelCase : str = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="DPT does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Tuple = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , nn.Linear ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : Optional[Any] = [*signature.parameters.keys()] UpperCamelCase : Optional[int] = ["pixel_values"] self.assertListEqual(arg_names[:1] , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase , UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : List[Any] = True if model_class in get_values(A_ ): continue UpperCamelCase : Optional[int] = model_class(A_ ) model.to(A_ ) model.train() UpperCamelCase : Optional[int] = self._prepare_for_class(A_ , A_ , return_labels=A_ ) UpperCamelCase : Optional[int] = model(**A_ ).loss loss.backward() def __UpperCamelCase( self ): '''simple docstring''' for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase , UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Any = False UpperCamelCase : Optional[int] = True if model_class in get_values(A_ ) or not model_class.supports_gradient_checkpointing: continue UpperCamelCase : str = model_class(A_ ) model.to(A_ ) model.gradient_checkpointing_enable() model.train() UpperCamelCase : Any = self._prepare_for_class(A_ , A_ , return_labels=A_ ) UpperCamelCase : Any = model(**A_ ).loss loss.backward() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : List[Any] = _config_zero_init(A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[str] = model_class(config=A_ ) # Skip the check for the backbone UpperCamelCase : List[Any] = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": UpperCamelCase : Optional[Any] = [F"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __UpperCamelCase( self ): '''simple docstring''' pass @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: UpperCamelCase : Dict = DPTModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = "add" with self.assertRaises(A_ ): UpperCamelCase : Optional[int] = DPTForDepthEstimation(A_ ) def A_ ( ) -> Optional[Any]: UpperCamelCase : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision @slow class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = DPTImageProcessor.from_pretrained("Intel/dpt-hybrid-midas" ) UpperCamelCase : Optional[Any] = DPTForDepthEstimation.from_pretrained("Intel/dpt-hybrid-midas" ).to(A_ ) UpperCamelCase : int = prepare_img() UpperCamelCase : Tuple = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : List[str] = model(**A_ ) UpperCamelCase : int = outputs.predicted_depth # verify the predicted depth UpperCamelCase : str = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , A_ ) UpperCamelCase : str = torch.tensor( [[[5.64_37, 5.61_46, 5.65_11], [5.43_71, 5.56_49, 5.59_58], [5.52_15, 5.51_84, 5.52_93]]] ).to(A_ ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , A_ , atol=1e-4 ) )
38
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A_ ( ) -> Dict: UpperCamelCase : Tuple = ArgumentParser( description=( "PyTorch TPU distributed training launch " "helper utility that will spawn up " "multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCAmelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCAmelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCAmelCase ) return parser.parse_args() def A_ ( ) -> Optional[int]: UpperCamelCase : Tuple = parse_args() # Import training_script as a module. UpperCamelCase : Union[str, Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase : List[Any] = script_fpath.stem UpperCamelCase : Optional[Any] = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv UpperCamelCase : List[Any] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
38
1
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __lowerCamelCase : int = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right __lowerCamelCase : List[Any] = 25_0004 __lowerCamelCase : List[Any] = 25_0020 @require_sentencepiece @require_tokenizers class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = MBartTokenizer _UpperCAmelCase :Tuple = MBartTokenizerFast _UpperCAmelCase :str = True _UpperCAmelCase :List[Any] = True def __UpperCamelCase( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase : List[str] = MBartTokenizer(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = MBartTokenizer(A_ , keep_accents=A_ ) UpperCamelCase : Tuple = tokenizer.tokenize("This is a test" ) self.assertListEqual(A_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCamelCase : Tuple = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) UpperCamelCase : List[Any] = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCamelCase : Tuple = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def __UpperCamelCase( self ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCamelCase : int = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCamelCase : Dict = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) UpperCamelCase : Any = self.tokenizer_class.from_pretrained(A_ , **A_ ) UpperCamelCase : Any = tempfile.mkdtemp() UpperCamelCase : Any = tokenizer_r.save_pretrained(A_ ) UpperCamelCase : Any = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) UpperCamelCase : Optional[Any] = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way UpperCamelCase : Union[str, Any] = tokenizer_r.from_pretrained(A_ ) UpperCamelCase : Any = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=True UpperCamelCase : Optional[Any] = tempfile.mkdtemp() UpperCamelCase : int = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) UpperCamelCase : Tuple = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way UpperCamelCase : int = tokenizer_r.from_pretrained(A_ ) UpperCamelCase : Dict = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=False UpperCamelCase : List[str] = tempfile.mkdtemp() UpperCamelCase : Optional[Any] = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) UpperCamelCase : Any = tokenizer_p.save_pretrained(A_ ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCamelCase : str = tokenizer_r.from_pretrained(A_ ) UpperCamelCase : Optional[int] = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) @require_torch @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase ): _UpperCAmelCase :Union[str, Any] = 'facebook/mbart-large-en-ro' _UpperCAmelCase :List[Any] = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] _UpperCAmelCase :Dict = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] _UpperCAmelCase :Optional[int] = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def __UpperCamelCase( cls ): '''simple docstring''' UpperCamelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO" ) UpperCamelCase : Any = 1 return cls def __UpperCamelCase( self ): '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 25_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 25_0004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 25_0020 ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A_ ) def __UpperCamelCase( self ): '''simple docstring''' self.assertIn(A_ , self.tokenizer.all_special_ids ) UpperCamelCase : List[Any] = [RO_CODE, 884, 9019, 96, 9, 916, 8_6792, 36, 1_8743, 1_5596, 5, 2] UpperCamelCase : Optional[Any] = self.tokenizer.decode(A_ , skip_special_tokens=A_ ) UpperCamelCase : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ ) self.assertNotIn(self.tokenizer.eos_token , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , A_ ) UpperCamelCase : Tuple = 10 UpperCamelCase : Optional[int] = self.tokenizer(A_ , max_length=A_ , truncation=A_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , A_ ) self.assertEqual(len(A_ ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"] ) , [25_0026, 25_0001] ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = tempfile.mkdtemp() UpperCamelCase : Dict = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(A_ ) UpperCamelCase : int = MBartTokenizer.from_pretrained(A_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , A_ ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=A_ , return_tensors="pt" ) UpperCamelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) UpperCamelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(A_ , A_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) UpperCamelCase : List[Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , A_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.tokenizer(self.src_text , padding=A_ , truncation=A_ , max_length=3 , return_tensors="pt" ) UpperCamelCase : List[Any] = self.tokenizer( text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=10 , return_tensors="pt" ) UpperCamelCase : List[str] = targets["input_ids"] UpperCamelCase : str = shift_tokens_right(A_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR" ) self.assertEqual( nested_simplify(A_ ) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 25_0004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 25_0001, } , )
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
1
import logging import os import threading import time try: import warnings except ImportError: __lowerCamelCase : str = None try: import msvcrt except ImportError: __lowerCamelCase : str = None try: import fcntl except ImportError: __lowerCamelCase : List[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __lowerCamelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ __lowerCamelCase : str = [ """Timeout""", """BaseFileLock""", """WindowsFileLock""", """UnixFileLock""", """SoftFileLock""", """FileLock""", ] __lowerCamelCase : Union[str, Any] = """3.0.12""" __lowerCamelCase : Any = None def A_ ( ) -> List[Any]: global _logger UpperCamelCase : Any = _logger or logging.getLogger(__name__ ) return _logger class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = lock_file return None def __str__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = lock return None def __enter__( self ): '''simple docstring''' return self.lock def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.lock.release() return None class A__ : def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : List[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long UpperCamelCase : Dict = self.hash_filename_if_too_long(A_ , A_ ) # The path to the lock file. UpperCamelCase : List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. UpperCamelCase : Tuple = None # The default timeout value. UpperCamelCase : Optional[Any] = timeout # We use this lock primarily for the lock counter. UpperCamelCase : Union[str, Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. UpperCamelCase : Dict = 0 return None @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file @property def __UpperCamelCase( self ): '''simple docstring''' return self._timeout @timeout.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = float(A_ ) return None def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file_fd is not None def __UpperCamelCase( self , A_=None , A_=0.05 ): '''simple docstring''' if timeout is None: UpperCamelCase : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 UpperCamelCase : Dict = id(self ) UpperCamelCase : List[str] = self._lock_file UpperCamelCase : int = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(A_ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: UpperCamelCase : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __UpperCamelCase( self , A_=False ): '''simple docstring''' with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: UpperCamelCase : List[Any] = id(self ) UpperCamelCase : Dict = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() UpperCamelCase : Dict = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): '''simple docstring''' self.acquire() return self def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.release() return None def __del__( self ): '''simple docstring''' self.release(force=A_ ) return None def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = os.path.basename(A_ ) if len(A_ ) > max_length and max_length > 0: UpperCamelCase : Optional[int] = os.path.dirname(A_ ) UpperCamelCase : int = str(hash(A_ ) ) UpperCamelCase : Any = filename[: max_length - len(A_ ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(A_ , A_ ) else: return path class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' from .file_utils import relative_to_absolute_path super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) UpperCamelCase : List[Any] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: UpperCamelCase : str = os.open(self._lock_file , A_ ) except OSError: pass else: try: msvcrt.locking(A_ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(A_ ) else: UpperCamelCase : Optional[Any] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self._lock_file_fd UpperCamelCase : str = None msvcrt.locking(A_ , msvcrt.LK_UNLCK , 1 ) os.close(A_ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : Tuple = os.statvfs(os.path.dirname(A_ ) ).f_namemax super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC UpperCamelCase : int = os.open(self._lock_file , A_ ) try: fcntl.flock(A_ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(A_ ) else: UpperCamelCase : List[str] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self._lock_file_fd UpperCamelCase : List[Any] = None fcntl.flock(A_ , fcntl.LOCK_UN ) os.close(A_ ) return None class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: UpperCamelCase : Optional[int] = os.open(self._lock_file , A_ ) except OSError: pass else: UpperCamelCase : Tuple = fd return None def __UpperCamelCase( self ): '''simple docstring''' os.close(self._lock_file_fd ) UpperCamelCase : str = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __lowerCamelCase : Dict = None if msvcrt: __lowerCamelCase : Any = WindowsFileLock elif fcntl: __lowerCamelCase : Any = UnixFileLock else: __lowerCamelCase : int = SoftFileLock if warnings is not None: warnings.warn("""only soft file lock is available""")
38
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __lowerCamelCase : Dict = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __lowerCamelCase : Tuple = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = VOCAB_FILES_NAMES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[Any] = BlenderbotSmallTokenizer def __init__( self , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , A_=True , **A_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=A_ , merges=A_ , add_prefix_space=A_ , trim_offsets=A_ , ) , bos_token=A_ , eos_token=A_ , unk_token=A_ , **A_ , ) UpperCamelCase : Union[str, Any] = add_prefix_space def __UpperCamelCase( self , A_ , A_=None ): '''simple docstring''' UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
38
1
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 __lowerCamelCase : Any = get_tests_dir("""fixtures""") class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = mock.Mock() UpperCamelCase : Any = 500 UpperCamelCase : Any = {} UpperCamelCase : Optional[Any] = HTTPError UpperCamelCase : str = {} # Download this model to make sure it's in the cache. UpperCamelCase : Optional[Any] = ViTImageProcessor.from_pretrained("hf-internal-testing/tiny-random-vit" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=A_ ) as mock_head: UpperCamelCase : Union[str, Any] = ViTImageProcessor.from_pretrained("hf-internal-testing/tiny-random-vit" ) # This check we did call the fake head request mock_head.assert_called() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = ViTImageProcessor.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json" ) def __UpperCamelCase( self ): '''simple docstring''' with self.assertRaises(A_ ): # config is in subfolder, the following should not work without specifying the subfolder UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained("hf-internal-testing/stable-diffusion-all-variants" ) UpperCamelCase : Any = AutoImageProcessor.from_pretrained( "hf-internal-testing/stable-diffusion-all-variants" , subfolder="feature_extractor" ) self.assertIsNotNone(A_ ) @is_staging_test class A__ ( unittest.TestCase ): @classmethod def __UpperCamelCase( cls ): '''simple docstring''' UpperCamelCase : Any = TOKEN HfFolder.save_token(A_ ) @classmethod def __UpperCamelCase( cls ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-image-processor" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-image-processor-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-image-processor" ) except HTTPError: pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = ViTImageProcessor.from_pretrained(A_ ) image_processor.push_to_hub("test-image-processor" , use_auth_token=self._token ) UpperCamelCase : int = ViTImageProcessor.from_pretrained(F"""{USER}/test-image-processor""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(A_ , getattr(A_ , A_ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-image-processor" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( A_ , repo_id="test-image-processor" , push_to_hub=A_ , use_auth_token=self._token ) UpperCamelCase : Union[str, Any] = ViTImageProcessor.from_pretrained(F"""{USER}/test-image-processor""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(A_ , getattr(A_ , A_ ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = ViTImageProcessor.from_pretrained(A_ ) image_processor.push_to_hub("valid_org/test-image-processor" , use_auth_token=self._token ) UpperCamelCase : Optional[int] = ViTImageProcessor.from_pretrained("valid_org/test-image-processor" ) for k, v in image_processor.__dict__.items(): self.assertEqual(A_ , getattr(A_ , A_ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-image-processor" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( A_ , repo_id="valid_org/test-image-processor-org" , push_to_hub=A_ , use_auth_token=self._token ) UpperCamelCase : Any = ViTImageProcessor.from_pretrained("valid_org/test-image-processor-org" ) for k, v in image_processor.__dict__.items(): self.assertEqual(A_ , getattr(A_ , A_ ) ) def __UpperCamelCase( self ): '''simple docstring''' CustomImageProcessor.register_for_auto_class() UpperCamelCase : str = CustomImageProcessor.from_pretrained(A_ ) image_processor.push_to_hub("test-dynamic-image-processor" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {"AutoImageProcessor": "custom_image_processing.CustomImageProcessor"} , ) UpperCamelCase : Any = AutoImageProcessor.from_pretrained( F"""{USER}/test-dynamic-image-processor""" , trust_remote_code=A_ ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , "CustomImageProcessor" )
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : int = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class A__ : _UpperCAmelCase :CommonSchedulerState # setable values _UpperCAmelCase :jnp.ndarray _UpperCAmelCase :jnp.ndarray _UpperCAmelCase :Optional[int] = None @classmethod def __UpperCamelCase( cls , A_ , A_ , A_ ): '''simple docstring''' return cls(common=A_ , init_noise_sigma=A_ , timesteps=A_ ) @dataclass class A__ ( __snake_case ): _UpperCAmelCase :DDPMSchedulerState class A__ ( __snake_case , __snake_case ): _UpperCAmelCase :Dict = [e.name for e in FlaxKarrasDiffusionSchedulers] _UpperCAmelCase :jnp.dtype @property def __UpperCamelCase( self ): '''simple docstring''' return True @register_to_config def __init__( self , A_ = 1000 , A_ = 0.00_01 , A_ = 0.02 , A_ = "linear" , A_ = None , A_ = "fixed_small" , A_ = True , A_ = "epsilon" , A_ = jnp.floataa , ): '''simple docstring''' UpperCamelCase : int = dtype def __UpperCamelCase( self , A_ = None ): '''simple docstring''' if common is None: UpperCamelCase : List[Any] = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution UpperCamelCase : List[Any] = jnp.array(1.0 , dtype=self.dtype ) UpperCamelCase : int = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=A_ , init_noise_sigma=A_ , timesteps=A_ , ) def __UpperCamelCase( self , A_ , A_ , A_ = None ): '''simple docstring''' return sample def __UpperCamelCase( self , A_ , A_ , A_ = () ): '''simple docstring''' UpperCamelCase : List[Any] = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 UpperCamelCase : Union[str, Any] = (jnp.arange(0 , A_ ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=A_ , timesteps=A_ , ) def __UpperCamelCase( self , A_ , A_ , A_=None , A_=None ): '''simple docstring''' UpperCamelCase : Optional[int] = state.common.alphas_cumprod[t] UpperCamelCase : List[str] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCamelCase : str = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: UpperCamelCase : str = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": UpperCamelCase : str = jnp.clip(A_ , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": UpperCamelCase : Optional[Any] = jnp.log(jnp.clip(A_ , a_min=1e-20 ) ) elif variance_type == "fixed_large": UpperCamelCase : Tuple = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log UpperCamelCase : Optional[int] = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": UpperCamelCase : Optional[int] = variance UpperCamelCase : Dict = state.common.betas[t] UpperCamelCase : Dict = (predicted_variance + 1) / 2 UpperCamelCase : List[Any] = frac * max_log + (1 - frac) * min_log return variance def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ = None , A_ = True , ): '''simple docstring''' UpperCamelCase : Union[str, Any] = timestep if key is None: UpperCamelCase : str = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: UpperCamelCase , UpperCamelCase : List[str] = jnp.split(A_ , sample.shape[1] , axis=1 ) else: UpperCamelCase : int = None # 1. compute alphas, betas UpperCamelCase : Tuple = state.common.alphas_cumprod[t] UpperCamelCase : List[str] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) UpperCamelCase : Union[str, Any] = 1 - alpha_prod_t UpperCamelCase : Tuple = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCamelCase : List[Any] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCamelCase : List[Any] = model_output elif self.config.prediction_type == "v_prediction": UpperCamelCase : Tuple = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` """ " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCamelCase : Dict = jnp.clip(A_ , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCamelCase : int = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t UpperCamelCase : Union[str, Any] = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCamelCase : Any = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): UpperCamelCase : Dict = jax.random.split(A_ , num=1 ) UpperCamelCase : List[Any] = jax.random.normal(A_ , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(A_ , A_ , predicted_variance=A_ ) ** 0.5) * noise UpperCamelCase : Tuple = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) UpperCamelCase : str = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=A_ , state=A_ ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , ): '''simple docstring''' return add_noise_common(state.common , A_ , A_ , A_ ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , ): '''simple docstring''' return get_velocity_common(state.common , A_ , A_ , A_ ) def __len__( self ): '''simple docstring''' return self.config.num_train_timesteps
38
import logging import os import threading import time try: import warnings except ImportError: __lowerCamelCase : str = None try: import msvcrt except ImportError: __lowerCamelCase : str = None try: import fcntl except ImportError: __lowerCamelCase : List[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __lowerCamelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ __lowerCamelCase : str = [ """Timeout""", """BaseFileLock""", """WindowsFileLock""", """UnixFileLock""", """SoftFileLock""", """FileLock""", ] __lowerCamelCase : Union[str, Any] = """3.0.12""" __lowerCamelCase : Any = None def A_ ( ) -> List[Any]: global _logger UpperCamelCase : Any = _logger or logging.getLogger(__name__ ) return _logger class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = lock_file return None def __str__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = lock return None def __enter__( self ): '''simple docstring''' return self.lock def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.lock.release() return None class A__ : def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : List[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long UpperCamelCase : Dict = self.hash_filename_if_too_long(A_ , A_ ) # The path to the lock file. UpperCamelCase : List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. UpperCamelCase : Tuple = None # The default timeout value. UpperCamelCase : Optional[Any] = timeout # We use this lock primarily for the lock counter. UpperCamelCase : Union[str, Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. UpperCamelCase : Dict = 0 return None @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file @property def __UpperCamelCase( self ): '''simple docstring''' return self._timeout @timeout.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = float(A_ ) return None def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file_fd is not None def __UpperCamelCase( self , A_=None , A_=0.05 ): '''simple docstring''' if timeout is None: UpperCamelCase : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 UpperCamelCase : Dict = id(self ) UpperCamelCase : List[str] = self._lock_file UpperCamelCase : int = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(A_ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: UpperCamelCase : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __UpperCamelCase( self , A_=False ): '''simple docstring''' with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: UpperCamelCase : List[Any] = id(self ) UpperCamelCase : Dict = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() UpperCamelCase : Dict = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): '''simple docstring''' self.acquire() return self def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.release() return None def __del__( self ): '''simple docstring''' self.release(force=A_ ) return None def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = os.path.basename(A_ ) if len(A_ ) > max_length and max_length > 0: UpperCamelCase : Optional[int] = os.path.dirname(A_ ) UpperCamelCase : int = str(hash(A_ ) ) UpperCamelCase : Any = filename[: max_length - len(A_ ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(A_ , A_ ) else: return path class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' from .file_utils import relative_to_absolute_path super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) UpperCamelCase : List[Any] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: UpperCamelCase : str = os.open(self._lock_file , A_ ) except OSError: pass else: try: msvcrt.locking(A_ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(A_ ) else: UpperCamelCase : Optional[Any] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self._lock_file_fd UpperCamelCase : str = None msvcrt.locking(A_ , msvcrt.LK_UNLCK , 1 ) os.close(A_ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : Tuple = os.statvfs(os.path.dirname(A_ ) ).f_namemax super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC UpperCamelCase : int = os.open(self._lock_file , A_ ) try: fcntl.flock(A_ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(A_ ) else: UpperCamelCase : List[str] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self._lock_file_fd UpperCamelCase : List[Any] = None fcntl.flock(A_ , fcntl.LOCK_UN ) os.close(A_ ) return None class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: UpperCamelCase : Optional[int] = os.open(self._lock_file , A_ ) except OSError: pass else: UpperCamelCase : Tuple = fd return None def __UpperCamelCase( self ): '''simple docstring''' os.close(self._lock_file_fd ) UpperCamelCase : str = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __lowerCamelCase : Dict = None if msvcrt: __lowerCamelCase : Any = WindowsFileLock elif fcntl: __lowerCamelCase : Any = UnixFileLock else: __lowerCamelCase : int = SoftFileLock if warnings is not None: warnings.warn("""only soft file lock is available""")
38
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : List[Any] = { """microsoft/resnet-50""": """https://huggingface.co/microsoft/resnet-50/blob/main/config.json""", } class A__ ( __snake_case , __snake_case ): _UpperCAmelCase :Optional[int] = 'resnet' _UpperCAmelCase :Optional[Any] = ['basic', 'bottleneck'] def __init__( self , A_=3 , A_=64 , A_=[256, 512, 1024, 2048] , A_=[3, 4, 6, 3] , A_="bottleneck" , A_="relu" , A_=False , A_=None , A_=None , **A_ , ): '''simple docstring''' super().__init__(**A_ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {",".join(self.layer_types )}""" ) UpperCamelCase : str = num_channels UpperCamelCase : Optional[int] = embedding_size UpperCamelCase : Dict = hidden_sizes UpperCamelCase : Union[str, Any] = depths UpperCamelCase : int = layer_type UpperCamelCase : str = hidden_act UpperCamelCase : List[Any] = downsample_in_first_stage UpperCamelCase : List[str] = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(A_ ) + 1 )] UpperCamelCase , UpperCamelCase : int = get_aligned_output_features_output_indices( out_features=A_ , out_indices=A_ , stage_names=self.stage_names ) class A__ ( __snake_case ): _UpperCAmelCase :List[str] = version.parse('1.11' ) @property def __UpperCamelCase( self ): '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def __UpperCamelCase( self ): '''simple docstring''' return 1e-3
38
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> str: if config_name_or_path is None: UpperCamelCase : Dict = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCamelCase : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCamelCase : Tuple = question_encoder_name_or_path UpperCamelCase : Any = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCamelCase : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Tuple = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : int = gen_config UpperCamelCase : Dict = question_encoder_config UpperCamelCase : Tuple = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--model_type""", choices=["""rag_sequence""", """rag_token"""], required=True, type=str, help="""RAG model type: rag_sequence, rag_token""", ) parser.add_argument("""--dest""", type=str, required=True, help="""Path to the output checkpoint directory.""") parser.add_argument("""--generator_name_or_path""", type=str, required=True, help="""Generator model identifier""") parser.add_argument( """--question_encoder_name_or_path""", type=str, required=True, help="""Question encoder model identifier""" ) parser.add_argument( """--generator_tokenizer_name_or_path""", type=str, help="""Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``""", ) parser.add_argument( """--question_encoder_tokenizer_name_or_path""", type=str, help="""Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``""", ) parser.add_argument( """--config_name_or_path""", type=str, help=( """Identifier of the model config to use, if not provided, resolves to a base config for a given""" """ ``model_type``""" ), ) __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
38
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class A__ : def __init__( self , A_ , A_=100 , A_=13 , A_=30 , A_=2 , A_=3 , A_=True , A_=True , A_=32 , A_=4 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=10 , A_=0.02 , A_=3 , A_=None , A_=[0, 1, 2, 3] , ): '''simple docstring''' UpperCamelCase : int = parent UpperCamelCase : Optional[Any] = 100 UpperCamelCase : List[str] = batch_size UpperCamelCase : int = image_size UpperCamelCase : str = patch_size UpperCamelCase : Optional[Any] = num_channels UpperCamelCase : str = is_training UpperCamelCase : Tuple = use_labels UpperCamelCase : int = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : Dict = intermediate_size UpperCamelCase : List[Any] = hidden_act UpperCamelCase : Union[str, Any] = hidden_dropout_prob UpperCamelCase : List[str] = attention_probs_dropout_prob UpperCamelCase : Dict = type_sequence_label_size UpperCamelCase : Union[str, Any] = initializer_range UpperCamelCase : int = scope UpperCamelCase : Optional[Any] = out_indices UpperCamelCase : Tuple = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCamelCase : List[str] = (image_size // patch_size) ** 2 UpperCamelCase : List[Any] = num_patches + 1 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : int = None UpperCamelCase : Optional[int] = None if self.use_labels: UpperCamelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : List[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase : int = self.get_config() return config, pixel_values, labels, pixel_labels def __UpperCamelCase( self ): '''simple docstring''' return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = BeitModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Any = BeitForMaskedImageModeling(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.type_sequence_label_size UpperCamelCase : List[Any] = BeitForImageClassification(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Tuple = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase : List[Any] = 1 UpperCamelCase : Any = BeitForImageClassification(A_ ) model.to(A_ ) model.eval() UpperCamelCase : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase : Tuple = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : List[str] = BeitForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() UpperCamelCase : int = model(A_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) UpperCamelCase : Tuple = model(A_ , labels=A_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : Dict = config_and_inputs UpperCamelCase : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) _UpperCAmelCase :List[str] = ( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :str = False _UpperCAmelCase :Optional[int] = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = BeitModelTester(self ) UpperCamelCase : List[Any] = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="BEiT does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason="BEiT has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Tuple = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , nn.Linear ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[Any] = model_class(A_ ) UpperCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : Optional[int] = [*signature.parameters.keys()] UpperCamelCase : List[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' if not self.model_tester.is_training: return UpperCamelCase , UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(A_ ), BeitForMaskedImageModeling]: continue UpperCamelCase : Dict = model_class(A_ ) model.to(A_ ) model.train() UpperCamelCase : List[str] = self._prepare_for_class(A_ , A_ , return_labels=A_ ) UpperCamelCase : int = model(**A_ ).loss loss.backward() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCamelCase : List[Any] = False UpperCamelCase : str = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(A_ ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue UpperCamelCase : Union[str, Any] = model_class(A_ ) model.gradient_checkpointing_enable() model.to(A_ ) model.train() UpperCamelCase : Optional[int] = self._prepare_for_class(A_ , A_ , return_labels=A_ ) UpperCamelCase : int = model(**A_ ).loss loss.backward() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : List[str] = _config_zero_init(A_ ) for model_class in self.all_model_classes: UpperCamelCase : Tuple = model_class(config=A_ ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Dict = BeitModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def A_ ( ) -> Optional[int]: UpperCamelCase : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def __UpperCamelCase( self ): '''simple docstring''' return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = BeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ).to(A_ ) UpperCamelCase : int = self.default_image_processor UpperCamelCase : int = prepare_img() UpperCamelCase : Optional[int] = image_processor(images=A_ , return_tensors="pt" ).pixel_values.to(A_ ) # prepare bool_masked_pos UpperCamelCase : Union[str, Any] = torch.ones((1, 196) , dtype=torch.bool ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Optional[int] = model(pixel_values=A_ , bool_masked_pos=A_ ) UpperCamelCase : Any = outputs.logits # verify the logits UpperCamelCase : List[str] = torch.Size((1, 196, 8192) ) self.assertEqual(logits.shape , A_ ) UpperCamelCase : List[Any] = torch.tensor( [[-3.24_37, 0.50_72, -13.91_74], [-3.24_56, 0.49_48, -13.94_01], [-3.20_33, 0.51_21, -13.85_50]] ).to(A_ ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , A_ , atol=1e-2 ) ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = BeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ).to(A_ ) UpperCamelCase : List[Any] = self.default_image_processor UpperCamelCase : Optional[Any] = prepare_img() UpperCamelCase : int = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Any = model(**A_ ) UpperCamelCase : List[str] = outputs.logits # verify the logits UpperCamelCase : Optional[int] = torch.Size((1, 1000) ) self.assertEqual(logits.shape , A_ ) UpperCamelCase : Optional[int] = torch.tensor([-1.23_85, -1.09_87, -1.01_08] ).to(A_ ) self.assertTrue(torch.allclose(logits[0, :3] , A_ , atol=1e-4 ) ) UpperCamelCase : Optional[int] = 281 self.assertEqual(logits.argmax(-1 ).item() , A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = BeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ).to( A_ ) UpperCamelCase : List[Any] = self.default_image_processor UpperCamelCase : Union[str, Any] = prepare_img() UpperCamelCase : Union[str, Any] = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Any = model(**A_ ) UpperCamelCase : Tuple = outputs.logits # verify the logits UpperCamelCase : str = torch.Size((1, 2_1841) ) self.assertEqual(logits.shape , A_ ) UpperCamelCase : List[Any] = torch.tensor([1.68_81, -0.27_87, 0.59_01] ).to(A_ ) self.assertTrue(torch.allclose(logits[0, :3] , A_ , atol=1e-4 ) ) UpperCamelCase : Optional[int] = 2396 self.assertEqual(logits.argmax(-1 ).item() , A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = BeitForSemanticSegmentation.from_pretrained("microsoft/beit-base-finetuned-ade-640-640" ) UpperCamelCase : Optional[Any] = model.to(A_ ) UpperCamelCase : Dict = BeitImageProcessor(do_resize=A_ , size=640 , do_center_crop=A_ ) UpperCamelCase : List[str] = load_dataset("hf-internal-testing/fixtures_ade20k" , split="test" ) UpperCamelCase : int = Image.open(ds[0]["file"] ) UpperCamelCase : Any = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : int = model(**A_ ) UpperCamelCase : Optional[int] = outputs.logits # verify the logits UpperCamelCase : int = torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape , A_ ) UpperCamelCase : Optional[Any] = version.parse(PIL.__version__ ) < version.parse("9.0.0" ) if is_pillow_less_than_a: UpperCamelCase : Tuple = torch.tensor( [ [[-4.92_25, -2.39_54, -3.05_22], [-2.88_22, -1.00_46, -1.75_61], [-2.95_49, -1.32_28, -2.13_47]], [[-5.81_68, -3.41_29, -4.07_78], [-3.86_51, -2.22_14, -3.02_77], [-3.83_56, -2.46_43, -3.35_35]], [[-0.00_78, 3.99_52, 4.07_54], [2.98_56, 4.69_44, 5.00_35], [3.24_13, 4.78_13, 4.99_69]], ] , device=A_ , ) else: UpperCamelCase : Optional[int] = torch.tensor( [ [[-4.89_60, -2.36_88, -3.03_55], [-2.84_78, -0.98_36, -1.74_18], [-2.94_49, -1.33_32, -2.14_56]], [[-5.80_81, -3.41_24, -4.10_06], [-3.85_61, -2.20_81, -3.03_23], [-3.83_65, -2.46_01, -3.36_69]], [[-0.03_09, 3.98_68, 4.05_40], [2.96_40, 4.68_77, 4.99_76], [3.20_81, 4.76_90, 4.99_42]], ] , device=A_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A_ , atol=1e-4 ) ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = BeitForSemanticSegmentation.from_pretrained("microsoft/beit-base-finetuned-ade-640-640" ) UpperCamelCase : List[Any] = model.to(A_ ) UpperCamelCase : str = BeitImageProcessor(do_resize=A_ , size=640 , do_center_crop=A_ ) UpperCamelCase : Optional[Any] = load_dataset("hf-internal-testing/fixtures_ade20k" , split="test" ) UpperCamelCase : List[Any] = Image.open(ds[0]["file"] ) UpperCamelCase : List[Any] = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : str = model(**A_ ) UpperCamelCase : List[Any] = outputs.logits.detach().cpu() UpperCamelCase : Optional[Any] = image_processor.post_process_semantic_segmentation(outputs=A_ , target_sizes=[(500, 300)] ) UpperCamelCase : int = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , A_ ) UpperCamelCase : Any = image_processor.post_process_semantic_segmentation(outputs=A_ ) UpperCamelCase : List[Any] = torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape , A_ )
38
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : str = 13 UpperCamelCase : int = 7 UpperCamelCase : str = True UpperCamelCase : Dict = True UpperCamelCase : str = True UpperCamelCase : Tuple = True UpperCamelCase : List[str] = 99 UpperCamelCase : Optional[Any] = 384 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Dict = 37 UpperCamelCase : Any = "gelu" UpperCamelCase : List[Any] = 0.1 UpperCamelCase : int = 0.1 UpperCamelCase : Tuple = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : int = 2 UpperCamelCase : Dict = 0.02 UpperCamelCase : Optional[Any] = 3 UpperCamelCase : List[Any] = 4 UpperCamelCase : Dict = 128 UpperCamelCase : Optional[Any] = 2 UpperCamelCase : Optional[int] = 9 UpperCamelCase : Optional[int] = 1 UpperCamelCase : Union[str, Any] = None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None if self.use_token_type_ids: UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel(config=A_ ) UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : Any = model(A_ ) UpperCamelCase : int = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = TFConvBertForMaskedLM(config=A_ ) UpperCamelCase : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = TFConvBertForSequenceClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : str = TFConvBertForMultipleChoice(config=A_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Dict = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Any = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : str = TFConvBertForTokenClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : str = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFConvBertForQuestionAnswering(config=A_ ) UpperCamelCase : Union[str, Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Union[str, Any] = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase :Any = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = TFConvBertModelTester(self ) UpperCamelCase : Dict = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Optional[Any] = True UpperCamelCase : Any = True if hasattr(A_ , "use_cache" ): UpperCamelCase : List[str] = True UpperCamelCase : List[Any] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Any = getattr(self.model_tester , "key_length" , A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[Any] = self._prepare_for_class(A_ , A_ ) UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Optional[int] = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) UpperCamelCase : Union[str, Any] = os.path.join(A_ , "saved_model" , "1" ) UpperCamelCase : Dict = tf.keras.models.load_model(A_ ) UpperCamelCase : str = model(A_ ) if self.is_encoder_decoder: UpperCamelCase : Union[str, Any] = outputs["encoder_hidden_states"] UpperCamelCase : Any = outputs["encoder_attentions"] else: UpperCamelCase : Any = outputs["hidden_states"] UpperCamelCase : List[str] = outputs["attentions"] self.assertEqual(len(A_ ) , A_ ) UpperCamelCase : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True UpperCamelCase : int = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "key_length" , A_ ) UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , A_ ) def check_decoder_attentions_output(A_ ): UpperCamelCase : Optional[Any] = len(A_ ) self.assertEqual(out_len % 2 , 0 ) UpperCamelCase : Any = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): UpperCamelCase : Dict = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True UpperCamelCase : List[Any] = False UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : List[str] = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Tuple = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase : Tuple = True UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine UpperCamelCase : Optional[int] = True UpperCamelCase : List[str] = True UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : Optional[Any] = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : List[str] = model(A_ )[0] UpperCamelCase : int = [1, 6, 768] self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
38
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = tempfile.mkdtemp() # fmt: off UpperCamelCase : List[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on UpperCamelCase : List[Any] = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase : List[Any] = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] UpperCamelCase : Any = {"unk_token": "<unk>"} UpperCamelCase : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(A_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(A_ ) ) UpperCamelCase : Dict = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], "image_std": [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], } UpperCamelCase : str = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(A_ , A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return CLIPTokenizer.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return ViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCamelCase : str = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.get_tokenizer() UpperCamelCase : Dict = self.get_rust_tokenizer() UpperCamelCase : List[Any] = self.get_image_processor() UpperCamelCase : Any = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase : int = CLIPSegProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) UpperCamelCase : str = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase : int = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , A_ ) self.assertIsInstance(processor_fast.tokenizer , A_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , A_ ) self.assertIsInstance(processor_fast.image_processor , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = CLIPSegProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase : Union[str, Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCamelCase : Tuple = self.get_image_processor(do_normalize=A_ , padding_value=1.0 ) UpperCamelCase : List[str] = CLIPSegProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=A_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.get_image_processor() UpperCamelCase : str = self.get_tokenizer() UpperCamelCase : List[Any] = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Union[str, Any] = self.prepare_image_inputs() UpperCamelCase : Optional[int] = image_processor(A_ , return_tensors="np" ) UpperCamelCase : str = processor(images=A_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.get_image_processor() UpperCamelCase : List[str] = self.get_tokenizer() UpperCamelCase : Tuple = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Tuple = "lower newer" UpperCamelCase : List[str] = processor(text=A_ ) UpperCamelCase : Optional[Any] = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.get_image_processor() UpperCamelCase : Tuple = self.get_tokenizer() UpperCamelCase : Union[str, Any] = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : int = "lower newer" UpperCamelCase : Any = self.prepare_image_inputs() UpperCamelCase : Optional[int] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.get_image_processor() UpperCamelCase : str = self.get_tokenizer() UpperCamelCase : List[Any] = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Optional[Any] = self.prepare_image_inputs() UpperCamelCase : Optional[Any] = self.prepare_image_inputs() UpperCamelCase : Optional[int] = processor(images=A_ , visual_prompt=A_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.get_image_processor() UpperCamelCase : List[Any] = self.get_tokenizer() UpperCamelCase : List[Any] = CLIPSegProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Optional[int] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase : Optional[int] = processor.batch_decode(A_ ) UpperCamelCase : Optional[int] = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ )
38
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : str = { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'camembert' def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : List[str] = vocab_size UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : Dict = hidden_act UpperCamelCase : str = intermediate_size UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : int = initializer_range UpperCamelCase : List[str] = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : int = use_cache UpperCamelCase : List[str] = classifier_dropout class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : str = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
38
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : str = logging.get_logger(__name__) __lowerCamelCase : str = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class A__ ( __snake_case ): _UpperCAmelCase :List[str] = 'openai-gpt' _UpperCAmelCase :Any = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , A_=4_0478 , A_=512 , A_=768 , A_=12 , A_=12 , A_="gelu" , A_=0.1 , A_=0.1 , A_=0.1 , A_=1e-5 , A_=0.02 , A_="cls_index" , A_=True , A_=None , A_=True , A_=0.1 , **A_ , ): '''simple docstring''' UpperCamelCase : Tuple = vocab_size UpperCamelCase : Tuple = n_positions UpperCamelCase : Tuple = n_embd UpperCamelCase : Tuple = n_layer UpperCamelCase : Union[str, Any] = n_head UpperCamelCase : int = afn UpperCamelCase : Dict = resid_pdrop UpperCamelCase : List[Any] = embd_pdrop UpperCamelCase : Tuple = attn_pdrop UpperCamelCase : Tuple = layer_norm_epsilon UpperCamelCase : Any = initializer_range UpperCamelCase : Dict = summary_type UpperCamelCase : Optional[Any] = summary_use_proj UpperCamelCase : int = summary_activation UpperCamelCase : Any = summary_first_dropout UpperCamelCase : Union[str, Any] = summary_proj_to_labels super().__init__(**A_ )
38
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: return int(input_a == input_a == 0 ) def A_ ( ) -> None: print("Truth Table of NOR Gate:" ) print("| Input 1 | Input 2 | Output |" ) print(F"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
from math import pow def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> tuple[int, int]: if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count UpperCamelCase : str = int(pow(_lowerCAmelCase , _lowerCAmelCase ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n UpperCamelCase , UpperCamelCase : int = backtrack( _lowerCAmelCase , _lowerCAmelCase , current_number + 1 , _lowerCAmelCase , _lowerCAmelCase ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. UpperCamelCase , UpperCamelCase : Optional[int] = backtrack( _lowerCAmelCase , _lowerCAmelCase , current_number + 1 , _lowerCAmelCase , _lowerCAmelCase ) return current_sum, solutions_count def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: if not (1 <= needed_sum <= 1000 and 2 <= power <= 10): raise ValueError( "Invalid input\n" "needed_sum must be between 1 and 1000, power between 2 and 10." ) return backtrack(_lowerCAmelCase , _lowerCAmelCase , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
38
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __snake_case ): _UpperCAmelCase :Optional[int] = ['image_processor', 'tokenizer'] _UpperCAmelCase :Tuple = 'BlipImageProcessor' _UpperCAmelCase :Optional[int] = 'AutoTokenizer' def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = False super().__init__(A_ , A_ ) UpperCamelCase : str = self.image_processor def __call__( self , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: UpperCamelCase : int = self.tokenizer UpperCamelCase : Optional[int] = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) return text_encoding # add pixel_values UpperCamelCase : int = self.image_processor(A_ , return_tensors=A_ ) if text is not None: UpperCamelCase : Dict = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) else: UpperCamelCase : Dict = None if text_encoding is not None: encoding_image_processor.update(A_ ) return encoding_image_processor def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.batch_decode(*A_ , **A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.decode(*A_ , **A_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.tokenizer.model_input_names UpperCamelCase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
38
1
def A_ ( _lowerCAmelCase ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence UpperCamelCase : Tuple = gray_code_sequence_string(_lowerCAmelCase ) # # convert them to integers for i in range(len(_lowerCAmelCase ) ): UpperCamelCase : Optional[int] = int(sequence[i] , 2 ) return sequence def A_ ( _lowerCAmelCase ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] UpperCamelCase : Tuple = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits UpperCamelCase : Any = gray_code_sequence_string(bit_count - 1 ) UpperCamelCase : List[str] = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): UpperCamelCase : Any = "0" + smaller_sequence[i] sequence.append(_lowerCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): UpperCamelCase : int = "1" + smaller_sequence[i] sequence.append(_lowerCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
38
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __lowerCamelCase : Dict = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Tuple = ['audio_values', 'audio_mask'] def __init__( self , A_=2048 , A_=1 , A_=[16, 16] , A_=128 , A_=4_4100 , A_=86 , A_=2048 , A_=0.0 , **A_ , ): '''simple docstring''' super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , **A_ , ) UpperCamelCase : Optional[int] = spectrogram_length UpperCamelCase : Dict = num_channels UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : str = feature_size // self.patch_size[1] UpperCamelCase : List[str] = n_fft UpperCamelCase : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase : Optional[int] = sampling_rate UpperCamelCase : int = padding_value UpperCamelCase : str = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=A_ , norm="slaney" , mel_scale="slaney" , ).T def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = spectrogram( A_ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) UpperCamelCase : List[Any] = log_spec[:, :-1] UpperCamelCase : Optional[int] = log_spec - 20.0 UpperCamelCase : str = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A_ , A_ = None , A_ = True , A_ = None , A_ = False , A_ = False , **A_ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase : Optional[int] = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase : Union[str, Any] = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): UpperCamelCase : str = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase : str = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A_ ): UpperCamelCase : int = [np.asarray(A_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase : Tuple = np.array(A_ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase : Union[str, Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase : Any = np.ones([len(A_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase : List[str] = padded_audio_features * self.padding_value for i in range(len(A_ ) ): UpperCamelCase : Union[str, Any] = audio_features[i] UpperCamelCase : Optional[int] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase : Optional[Any] = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: UpperCamelCase : int = {"audio_values": padded_audio_features} UpperCamelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
38
1
def A_ ( _lowerCAmelCase ) -> bool: UpperCamelCase : Optional[int] = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
38
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Dict = TypeVar("""KT""") __lowerCamelCase : Dict = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self , A_ = "root" , A_ = None ): '''simple docstring''' UpperCamelCase : int = key UpperCamelCase : List[Any] = value UpperCamelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F"""Node({self.key}: {self.value})""" @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self , A_ = 0.5 , A_ = 16 ): '''simple docstring''' UpperCamelCase : Node[KT, VT] = Node[KT, VT]() UpperCamelCase : List[Any] = 0 UpperCamelCase : Union[str, Any] = p UpperCamelCase : List[str] = max_level def __str__( self ): '''simple docstring''' UpperCamelCase : int = list(self ) if len(A_ ) == 0: return F"""SkipList(level={self.level})""" UpperCamelCase : str = max((len(str(A_ ) ) for item in items) , default=4 ) UpperCamelCase : Dict = max(A_ , 4 ) + 4 UpperCamelCase : str = self.head UpperCamelCase : List[Any] = [] UpperCamelCase : int = node.forward.copy() lines.append(F"""[{node.key}]""".ljust(A_ , "-" ) + "* " * len(A_ ) ) lines.append(" " * label_size + "| " * len(A_ ) ) while len(node.forward ) != 0: UpperCamelCase : Union[str, Any] = node.forward[0] lines.append( F"""[{node.key}]""".ljust(A_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(A_ ) ) UpperCamelCase : Tuple = node.forward lines.append("None".ljust(A_ ) + "* " * len(A_ ) ) return F"""SkipList(level={self.level})\n""" + "\n".join(A_ ) def __iter__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCamelCase : Union[str, Any] = node.forward[0] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = 1 while random() < self.p and level < self.max_level: level += 1 return level def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = [] UpperCamelCase : List[Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCamelCase : str = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(A_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self._locate_node(A_ ) if node is not None: for i, update_node in enumerate(A_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCamelCase : Tuple = node.forward[i] else: UpperCamelCase : List[Any] = update_node.forward[:i] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self._locate_node(A_ ) if node is not None: UpperCamelCase : Union[str, Any] = value else: UpperCamelCase : Dict = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , A_ ): update_vector.append(self.head ) UpperCamelCase : Optional[int] = level UpperCamelCase : Dict = Node(A_ , A_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(A_ ) else: UpperCamelCase : List[Any] = new_node def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self._locate_node(A_ ) if node is not None: return node.value return None def A_ ( ) -> List[Any]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) UpperCamelCase : Optional[int] = skip_list.head UpperCamelCase : List[str] = {} while node.level != 0: UpperCamelCase : str = node.forward[0] UpperCamelCase : Optional[int] = node.value assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def A_ ( ) -> List[Any]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) UpperCamelCase : Dict = skip_list.head UpperCamelCase : Tuple = {} while node.level != 0: UpperCamelCase : List[str] = node.forward[0] UpperCamelCase : Dict = node.value if len(_lowerCAmelCase ) != 4: print() assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() assert skip_list.find("Some key" ) is None def A_ ( ) -> Tuple: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def A_ ( ) -> Dict: UpperCamelCase : Optional[int] = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def A_ ( ) -> Dict: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[str]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_lowerCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_lowerCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def A_ ( ) -> Union[str, Any]: def is_sorted(_lowerCAmelCase ): return all(next_item >= item for item, next_item in zip(_lowerCAmelCase , lst[1:] ) ) UpperCamelCase : int = SkipList() for i in range(10 ): skip_list.insert(_lowerCAmelCase , _lowerCAmelCase ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_lowerCAmelCase ) ) def A_ ( ) -> Tuple: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def A_ ( ) -> List[str]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
__lowerCamelCase : dict[str, float] = { "joule": 1.0, "kilojoule": 1000, "megajoule": 100_0000, "gigajoule": 10_0000_0000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 360_0000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 418_6800.00, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1055.0_5585, "footpound": 1.3_5_5_8_1_8, } def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> float: if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: UpperCamelCase : List[Any] = ( F"""Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n""" F"""Valid values are: {", ".join(_lowerCAmelCase )}""" ) raise ValueError(_lowerCAmelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
38
from PIL import Image def A_ ( _lowerCAmelCase ) -> Image: UpperCamelCase , UpperCamelCase : List[Any] = image.size UpperCamelCase : Union[str, Any] = 0 UpperCamelCase : List[str] = image.load() for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(_lowerCAmelCase ): for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
38
1
import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask __lowerCamelCase : Dict = logging.getLogger(__name__) class A__ ( __snake_case ): def __init__( self , A_=-1 ): '''simple docstring''' UpperCamelCase : Optional[Any] = label_idx def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if isinstance(A_ , A_ ): UpperCamelCase : str = mode.value UpperCamelCase : List[Any] = os.path.join(A_ , F"""{mode}.txt""" ) UpperCamelCase : Union[str, Any] = 1 UpperCamelCase : List[Any] = [] with open(A_ , encoding="utf-8" ) as f: UpperCamelCase : Dict = [] UpperCamelCase : Optional[int] = [] for line in f: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=A_ , labels=A_ ) ) guid_index += 1 UpperCamelCase : Optional[int] = [] UpperCamelCase : List[Any] = [] else: UpperCamelCase : Dict = line.split(" " ) words.append(splits[0] ) if len(A_ ) > 1: labels.append(splits[self.label_idx].replace("\n" , "" ) ) else: # Examples could have no label for mode = "test" labels.append("O" ) if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=A_ , labels=A_ ) ) return examples def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = 0 for line in test_input_reader: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": writer.write(A_ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: UpperCamelCase : Optional[int] = line.split()[0] + " " + preds_list[example_id].pop(0 ) + "\n" writer.write(A_ ) else: logger.warning("Maximum sequence length exceeded: No prediction for '%s'." , line.split()[0] ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if path: with open(A_ , "r" ) as f: UpperCamelCase : Union[str, Any] = f.read().splitlines() if "O" not in labels: UpperCamelCase : Union[str, Any] = ["O"] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class A__ ( __snake_case ): def __init__( self ): '''simple docstring''' super().__init__(label_idx=-2 ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if path: with open(A_ , "r" ) as f: UpperCamelCase : List[str] = f.read().splitlines() if "O" not in labels: UpperCamelCase : Tuple = ["O"] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class A__ ( __snake_case ): def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if isinstance(A_ , A_ ): UpperCamelCase : List[str] = mode.value UpperCamelCase : int = os.path.join(A_ , F"""{mode}.txt""" ) UpperCamelCase : Dict = 1 UpperCamelCase : Optional[Any] = [] with open(A_ , encoding="utf-8" ) as f: for sentence in parse_incr(A_ ): UpperCamelCase : int = [] UpperCamelCase : Tuple = [] for token in sentence: words.append(token["form"] ) labels.append(token["upos"] ) assert len(A_ ) == len(A_ ) if words: examples.append(InputExample(guid=F"""{mode}-{guid_index}""" , words=A_ , labels=A_ ) ) guid_index += 1 return examples def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = 0 for sentence in parse_incr(A_ ): UpperCamelCase : Optional[Any] = preds_list[example_id] UpperCamelCase : str = "" for token in sentence: out += F"""{token["form"]} ({token["upos"]}|{s_p.pop(0 )}) """ out += "\n" writer.write(A_ ) example_id += 1 def __UpperCamelCase( self , A_ ): '''simple docstring''' if path: with open(A_ , "r" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
38
from math import loga def A_ ( _lowerCAmelCase ) -> int: if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
1
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class A__ ( __snake_case ): _UpperCAmelCase :jnp.ndarray _UpperCAmelCase :jnp.ndarray class A__ ( nn.Module ): _UpperCAmelCase :int _UpperCAmelCase :Tuple[int] = (1_6, 3_2, 9_6, 2_5_6) _UpperCAmelCase :jnp.dtype = jnp.floataa def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCamelCase : Optional[int] = [] for i in range(len(self.block_out_channels ) - 1 ): UpperCamelCase : List[Any] = self.block_out_channels[i] UpperCamelCase : Tuple = self.block_out_channels[i + 1] UpperCamelCase : List[str] = nn.Conv( A_ , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(A_ ) UpperCamelCase : List[str] = nn.Conv( A_ , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(A_ ) UpperCamelCase : List[Any] = blocks UpperCamelCase : Dict = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.conv_in(A_ ) UpperCamelCase : Any = nn.silu(A_ ) for block in self.blocks: UpperCamelCase : int = block(A_ ) UpperCamelCase : Any = nn.silu(A_ ) UpperCamelCase : List[Any] = self.conv_out(A_ ) return embedding @flax_register_to_config class A__ ( nn.Module , __snake_case , __snake_case ): _UpperCAmelCase :int = 3_2 _UpperCAmelCase :int = 4 _UpperCAmelCase :Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) _UpperCAmelCase :Union[bool, Tuple[bool]] = False _UpperCAmelCase :Tuple[int] = (3_2_0, 6_4_0, 1_2_8_0, 1_2_8_0) _UpperCAmelCase :int = 2 _UpperCAmelCase :Union[int, Tuple[int]] = 8 _UpperCAmelCase :Optional[Union[int, Tuple[int]]] = None _UpperCAmelCase :int = 1_2_8_0 _UpperCAmelCase :float = 0.0 _UpperCAmelCase :bool = False _UpperCAmelCase :jnp.dtype = jnp.floataa _UpperCAmelCase :bool = True _UpperCAmelCase :int = 0 _UpperCAmelCase :str = "rgb" _UpperCAmelCase :Tuple[int] = (1_6, 3_2, 9_6, 2_5_6) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = (1, self.in_channels, self.sample_size, self.sample_size) UpperCamelCase : Any = jnp.zeros(A_ , dtype=jnp.floataa ) UpperCamelCase : Union[str, Any] = jnp.ones((1,) , dtype=jnp.intaa ) UpperCamelCase : Dict = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) UpperCamelCase : List[str] = (1, 3, self.sample_size * 8, self.sample_size * 8) UpperCamelCase : Optional[Any] = jnp.zeros(A_ , dtype=jnp.floataa ) UpperCamelCase , UpperCamelCase : Union[str, Any] = jax.random.split(A_ ) UpperCamelCase : List[str] = {"params": params_rng, "dropout": dropout_rng} return self.init(A_ , A_ , A_ , A_ , A_ )["params"] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.block_out_channels UpperCamelCase : Optional[Any] = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCamelCase : Dict = self.num_attention_heads or self.attention_head_dim # input UpperCamelCase : int = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time UpperCamelCase : Optional[Any] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) UpperCamelCase : int = FlaxTimestepEmbedding(A_ , dtype=self.dtype ) UpperCamelCase : int = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) UpperCamelCase : Tuple = self.only_cross_attention if isinstance(A_ , A_ ): UpperCamelCase : Dict = (only_cross_attention,) * len(self.down_block_types ) if isinstance(A_ , A_ ): UpperCamelCase : Any = (num_attention_heads,) * len(self.down_block_types ) # down UpperCamelCase : str = [] UpperCamelCase : Optional[Any] = [] UpperCamelCase : int = block_out_channels[0] UpperCamelCase : str = nn.Conv( A_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(A_ ) for i, down_block_type in enumerate(self.down_block_types ): UpperCamelCase : Union[str, Any] = output_channel UpperCamelCase : Optional[Any] = block_out_channels[i] UpperCamelCase : int = i == len(A_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCamelCase : List[Any] = FlaxCrossAttnDownBlockaD( in_channels=A_ , out_channels=A_ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: UpperCamelCase : Any = FlaxDownBlockaD( in_channels=A_ , out_channels=A_ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(A_ ) for _ in range(self.layers_per_block ): UpperCamelCase : List[Any] = nn.Conv( A_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(A_ ) if not is_final_block: UpperCamelCase : int = nn.Conv( A_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(A_ ) UpperCamelCase : List[str] = down_blocks UpperCamelCase : List[str] = controlnet_down_blocks # mid UpperCamelCase : int = block_out_channels[-1] UpperCamelCase : Any = FlaxUNetMidBlockaDCrossAttn( in_channels=A_ , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) UpperCamelCase : Dict = nn.Conv( A_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , A_ , A_ , A_ , A_ , A_ = 1.0 , A_ = True , A_ = False , ): '''simple docstring''' UpperCamelCase : Tuple = self.controlnet_conditioning_channel_order if channel_order == "bgr": UpperCamelCase : str = jnp.flip(A_ , axis=1 ) # 1. time if not isinstance(A_ , jnp.ndarray ): UpperCamelCase : Tuple = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(A_ , jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCamelCase : Optional[Any] = timesteps.astype(dtype=jnp.floataa ) UpperCamelCase : List[Any] = jnp.expand_dims(A_ , 0 ) UpperCamelCase : int = self.time_proj(A_ ) UpperCamelCase : List[Any] = self.time_embedding(A_ ) # 2. pre-process UpperCamelCase : int = jnp.transpose(A_ , (0, 2, 3, 1) ) UpperCamelCase : Dict = self.conv_in(A_ ) UpperCamelCase : List[Any] = jnp.transpose(A_ , (0, 2, 3, 1) ) UpperCamelCase : Optional[int] = self.controlnet_cond_embedding(A_ ) sample += controlnet_cond # 3. down UpperCamelCase : Any = (sample,) for down_block in self.down_blocks: if isinstance(A_ , A_ ): UpperCamelCase , UpperCamelCase : Union[str, Any] = down_block(A_ , A_ , A_ , deterministic=not train ) else: UpperCamelCase , UpperCamelCase : Any = down_block(A_ , A_ , deterministic=not train ) down_block_res_samples += res_samples # 4. mid UpperCamelCase : Any = self.mid_block(A_ , A_ , A_ , deterministic=not train ) # 5. contronet blocks UpperCamelCase : List[Any] = () for down_block_res_sample, controlnet_block in zip(A_ , self.controlnet_down_blocks ): UpperCamelCase : Tuple = controlnet_block(A_ ) controlnet_down_block_res_samples += (down_block_res_sample,) UpperCamelCase : int = controlnet_down_block_res_samples UpperCamelCase : Optional[int] = self.controlnet_mid_block(A_ ) # 6. scaling UpperCamelCase : str = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=A_ , mid_block_res_sample=A_ )
38
from __future__ import annotations __lowerCamelCase : Optional[int] = """Muhammad Umer Farooq""" __lowerCamelCase : Tuple = """MIT""" __lowerCamelCase : Optional[int] = """1.0.0""" __lowerCamelCase : int = """Muhammad Umer Farooq""" __lowerCamelCase : Optional[int] = """contact@muhammadumerfarooq.me""" __lowerCamelCase : Dict = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : list[str] = [] UpperCamelCase : str = domain def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCamelCase : Any = parse.urljoin(self.domain , A_ ) self.urls.append(A_ ) def A_ ( _lowerCAmelCase ) -> str: return ".".join(get_sub_domain_name(_lowerCAmelCase ).split("." )[-2:] ) def A_ ( _lowerCAmelCase ) -> str: return parse.urlparse(_lowerCAmelCase ).netloc def A_ ( _lowerCAmelCase = "https://github.com" ) -> list[str]: UpperCamelCase : int = get_domain_name(_lowerCAmelCase ) # Initialize the parser UpperCamelCase : str = Parser(_lowerCAmelCase ) try: # Open URL UpperCamelCase : int = requests.get(_lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCamelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCamelCase : Optional[Any] = requests.get(_lowerCAmelCase ) # Get the valid email. UpperCamelCase : Optional[int] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = emails_from_url("""https://github.com""") print(f"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
38
1
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __lowerCamelCase : Dict = logging.getLogger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Any = 'sequence-classification' def __init__( self , A_ ): '''simple docstring''' if type(A_ ) == dict: UpperCamelCase : Any = Namespace(**A_ ) UpperCamelCase : Optional[int] = glue_output_modes[hparams.task] UpperCamelCase : Any = glue_tasks_num_labels[hparams.task] super().__init__(A_ , A_ , self.mode ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return self.model(**A_ ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Any = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: UpperCamelCase : Optional[int] = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None UpperCamelCase : List[Any] = self(**A_ ) UpperCamelCase : Dict = outputs[0] UpperCamelCase : Optional[Any] = self.trainer.lr_schedulers[0]["scheduler"] UpperCamelCase : int = {"loss": loss, "rate": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.hparams UpperCamelCase : Tuple = processors[args.task]() UpperCamelCase : List[str] = processor.get_labels() for mode in ["train", "dev"]: UpperCamelCase : str = self._feature_file(A_ ) if os.path.exists(A_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , A_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) UpperCamelCase : Optional[Any] = ( processor.get_dev_examples(args.data_dir ) if mode == "dev" else processor.get_train_examples(args.data_dir ) ) UpperCamelCase : Any = convert_examples_to_features( A_ , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("Saving features into cached file %s" , A_ ) torch.save(A_ , A_ ) def __UpperCamelCase( self , A_ , A_ , A_ = False ): '''simple docstring''' UpperCamelCase : Dict = "dev" if mode == "test" else mode UpperCamelCase : str = self._feature_file(A_ ) logger.info("Loading features from cached file %s" , A_ ) UpperCamelCase : Dict = torch.load(A_ ) UpperCamelCase : List[Any] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) UpperCamelCase : Tuple = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) UpperCamelCase : Optional[Any] = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": UpperCamelCase : List[str] = torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": UpperCamelCase : List[str] = torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(A_ , A_ , A_ , A_ ) , batch_size=A_ , shuffle=A_ , ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: UpperCamelCase : Union[str, Any] = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None UpperCamelCase : Tuple = self(**A_ ) UpperCamelCase , UpperCamelCase : Dict = outputs[:2] UpperCamelCase : Dict = logits.detach().cpu().numpy() UpperCamelCase : int = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Tuple = torch.stack([x["val_loss"] for x in outputs] ).mean().detach().cpu().item() UpperCamelCase : Optional[int] = np.concatenate([x["pred"] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": UpperCamelCase : Any = np.argmax(A_ , axis=1 ) elif self.hparams.glue_output_mode == "regression": UpperCamelCase : List[str] = np.squeeze(A_ ) UpperCamelCase : Union[str, Any] = np.concatenate([x["target"] for x in outputs] , axis=0 ) UpperCamelCase : List[str] = [[] for _ in range(out_label_ids.shape[0] )] UpperCamelCase : Any = [[] for _ in range(out_label_ids.shape[0] )] UpperCamelCase : List[str] = {**{"val_loss": val_loss_mean}, **compute_metrics(self.hparams.task , A_ , A_ )} UpperCamelCase : Optional[Any] = dict(results.items() ) UpperCamelCase : str = results return ret, preds_list, out_label_list def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase : List[Any] = self._eval_end(A_ ) UpperCamelCase : Union[str, Any] = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase : List[Any] = self._eval_end(A_ ) UpperCamelCase : List[Any] = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def __UpperCamelCase( A_ , A_ ): '''simple docstring''' BaseTransformer.add_model_specific_args(A_ , A_ ) parser.add_argument( "--max_seq_length" , default=128 , type=A_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--task" , default="" , type=A_ , required=A_ , help="The GLUE task to run" , ) parser.add_argument( "--gpus" , default=0 , type=A_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser def A_ ( ) -> Optional[int]: UpperCamelCase : Dict = argparse.ArgumentParser() add_generic_args(_lowerCAmelCase , os.getcwd() ) UpperCamelCase : Optional[int] = GLUETransformer.add_model_specific_args(_lowerCAmelCase , os.getcwd() ) UpperCamelCase : List[str] = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: UpperCamelCase : List[Any] = os.path.join( "./results" , F"""{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}""" , ) os.makedirs(args.output_dir ) UpperCamelCase : Optional[int] = GLUETransformer(_lowerCAmelCase ) UpperCamelCase : str = generic_train(_lowerCAmelCase , _lowerCAmelCase ) # Optionally, predict on dev set and write to output_dir if args.do_predict: UpperCamelCase : List[str] = sorted(glob.glob(os.path.join(args.output_dir , "checkpoint-epoch=*.ckpt" ) , recursive=_lowerCAmelCase ) ) UpperCamelCase : Tuple = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(_lowerCAmelCase ) if __name__ == "__main__": main()
38
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> list[int]: UpperCamelCase : Optional[Any] = [True] * limit UpperCamelCase : Optional[Any] = False UpperCamelCase : List[str] = False UpperCamelCase : Tuple = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCamelCase : Optional[Any] = i * 2 while index < limit: UpperCamelCase : int = False UpperCamelCase : Optional[int] = index + i UpperCamelCase : Any = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def A_ ( _lowerCAmelCase = 100_0000 ) -> int: UpperCamelCase : Union[str, Any] = prime_sieve(_lowerCAmelCase ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCamelCase : Dict = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCamelCase : int = j - i UpperCamelCase : Dict = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
38
1
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __snake_case ): def __init__( self , A_ , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , A_ = None , **A_ , ): '''simple docstring''' super().__init__( features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , num_proc=A_ , **A_ , ) UpperCamelCase : Optional[int] = Generator( cache_dir=A_ , features=A_ , generator=A_ , gen_kwargs=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' if self.streaming: UpperCamelCase : Optional[Any] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: UpperCamelCase : Union[str, Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , num_proc=self.num_proc , ) UpperCamelCase : int = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset
38
1
import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> List[str]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> int: UpperCamelCase : List[str] = tmp_path / "cache" UpperCamelCase : Tuple = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : List[Any] = TextDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_text_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[int]: UpperCamelCase : Union[str, Any] = tmp_path / "cache" UpperCamelCase : Tuple = {"text": "string"} UpperCamelCase : List[Any] = features.copy() if features else default_expected_features UpperCamelCase : Optional[int] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : Union[str, Any] = TextDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_text_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Any: UpperCamelCase : Any = tmp_path / "cache" UpperCamelCase : str = {"text": "string"} UpperCamelCase : Union[str, Any] = TextDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_text_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[str]: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCamelCase : List[str] = text_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCamelCase : Optional[int] = [text_path] UpperCamelCase : int = tmp_path / "cache" UpperCamelCase : Tuple = {"text": "string"} UpperCamelCase : Optional[int] = TextDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_text_dataset(_lowerCAmelCase , _lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=("train",) ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCamelCase : Union[str, Any] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Dict = tmp_path / "cache" UpperCamelCase : Tuple = {"text": "string"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : Any = TextDatasetReader({"train": text_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_text_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( "features" , [ None, {"text": "string"}, {"text": "int32"}, {"text": "float32"}, ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> int: UpperCamelCase : Any = tmp_path / "cache" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" UpperCamelCase : Dict = {"text": "string"} UpperCamelCase : Any = features.copy() if features else default_expected_features UpperCamelCase : int = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : str = TextDatasetReader({"train": text_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_text_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Dict: if split: UpperCamelCase : Optional[int] = {split: text_path} else: UpperCamelCase : str = "train" UpperCamelCase : Dict = {"train": text_path, "test": text_path} UpperCamelCase : Dict = tmp_path / "cache" UpperCamelCase : List[str] = {"text": "string"} UpperCamelCase : Optional[int] = TextDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_text_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
38
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __snake_case ): _UpperCAmelCase :Any = (PNDMScheduler,) _UpperCAmelCase :List[str] = (('num_inference_steps', 5_0),) def __UpperCamelCase( self , **A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**A_ ) return config def __UpperCamelCase( self , A_=0 , **A_ ): '''simple docstring''' UpperCamelCase : Any = dict(self.forward_default_kwargs ) UpperCamelCase : Any = kwargs.pop("num_inference_steps" , A_ ) UpperCamelCase : Optional[Any] = self.dummy_sample UpperCamelCase : str = 0.1 * sample UpperCamelCase : str = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCamelCase : Tuple = self.get_scheduler_config(**A_ ) UpperCamelCase : Dict = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals UpperCamelCase : List[str] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) UpperCamelCase : List[str] = scheduler_class.from_pretrained(A_ ) new_scheduler.set_timesteps(A_ ) # copy over dummy past residuals UpperCamelCase : List[Any] = dummy_past_residuals[:] UpperCamelCase : int = scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : Tuple = new_scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCamelCase : str = scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : List[Any] = new_scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self , A_=0 , **A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = dict(self.forward_default_kwargs ) UpperCamelCase : Union[str, Any] = kwargs.pop("num_inference_steps" , A_ ) UpperCamelCase : str = self.dummy_sample UpperCamelCase : Dict = 0.1 * sample UpperCamelCase : Union[str, Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCamelCase : Dict = self.get_scheduler_config() UpperCamelCase : Optional[int] = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCamelCase : Dict = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) UpperCamelCase : Dict = scheduler_class.from_pretrained(A_ ) # copy over dummy past residuals new_scheduler.set_timesteps(A_ ) # copy over dummy past residual (must be after setting timesteps) UpperCamelCase : Dict = dummy_past_residuals[:] UpperCamelCase : Optional[Any] = scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : Union[str, Any] = new_scheduler.step_prk(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCamelCase : Any = scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : int = new_scheduler.step_plms(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase( self , **A_ ): '''simple docstring''' UpperCamelCase : Any = self.scheduler_classes[0] UpperCamelCase : str = self.get_scheduler_config(**A_ ) UpperCamelCase : Dict = scheduler_class(**A_ ) UpperCamelCase : Dict = 10 UpperCamelCase : Any = self.dummy_model() UpperCamelCase : int = self.dummy_sample_deter scheduler.set_timesteps(A_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCamelCase : Dict = model(A_ , A_ ) UpperCamelCase : Tuple = scheduler.step_prk(A_ , A_ , A_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCamelCase : List[str] = model(A_ , A_ ) UpperCamelCase : Dict = scheduler.step_plms(A_ , A_ , A_ ).prev_sample return sample def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = dict(self.forward_default_kwargs ) UpperCamelCase : Dict = kwargs.pop("num_inference_steps" , A_ ) for scheduler_class in self.scheduler_classes: UpperCamelCase : int = self.get_scheduler_config() UpperCamelCase : Dict = scheduler_class(**A_ ) UpperCamelCase : Any = self.dummy_sample UpperCamelCase : Any = 0.1 * sample if num_inference_steps is not None and hasattr(A_ , "set_timesteps" ): scheduler.set_timesteps(A_ ) elif num_inference_steps is not None and not hasattr(A_ , "set_timesteps" ): UpperCamelCase : Optional[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCamelCase : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCamelCase : Optional[int] = dummy_past_residuals[:] UpperCamelCase : Tuple = scheduler.step_prk(A_ , 0 , A_ , **A_ ).prev_sample UpperCamelCase : Any = scheduler.step_prk(A_ , 1 , A_ , **A_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCamelCase : Dict = scheduler.step_plms(A_ , 0 , A_ , **A_ ).prev_sample UpperCamelCase : Tuple = scheduler.step_plms(A_ , 1 , A_ , **A_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __UpperCamelCase( self ): '''simple docstring''' for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for steps_offset in [0, 1]: self.check_over_configs(steps_offset=A_ ) UpperCamelCase : Any = self.scheduler_classes[0] UpperCamelCase : Any = self.get_scheduler_config(steps_offset=1 ) UpperCamelCase : List[Any] = scheduler_class(**A_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def __UpperCamelCase( self ): '''simple docstring''' for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=A_ , beta_end=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for t in [1, 5, 10]: self.check_over_forward(time_step=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCamelCase : str = self.dummy_sample UpperCamelCase : List[Any] = 0.1 * sample UpperCamelCase : Optional[Any] = self.get_scheduler_config() UpperCamelCase : List[str] = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCamelCase : Dict = scheduler.step_prk(A_ , A_ , A_ ).prev_sample def __UpperCamelCase( self ): '''simple docstring''' with self.assertRaises(A_ ): UpperCamelCase : Tuple = self.scheduler_classes[0] UpperCamelCase : Any = self.get_scheduler_config() UpperCamelCase : int = scheduler_class(**A_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.full_loop() UpperCamelCase : Dict = torch.sum(torch.abs(A_ ) ) UpperCamelCase : str = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1e-2 assert abs(result_mean.item() - 0.25_80 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.full_loop(prediction_type="v_prediction" ) UpperCamelCase : Optional[int] = torch.sum(torch.abs(A_ ) ) UpperCamelCase : List[str] = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1e-2 assert abs(result_mean.item() - 0.08_78 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.full_loop(set_alpha_to_one=A_ , beta_start=0.01 ) UpperCamelCase : Any = torch.sum(torch.abs(A_ ) ) UpperCamelCase : Tuple = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1e-2 assert abs(result_mean.item() - 0.29_95 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.full_loop(set_alpha_to_one=A_ , beta_start=0.01 ) UpperCamelCase : Optional[Any] = torch.sum(torch.abs(A_ ) ) UpperCamelCase : Optional[int] = torch.mean(torch.abs(A_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1e-2 assert abs(result_mean.item() - 0.24_34 ) < 1e-3
38
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
1
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __lowerCamelCase : Union[str, Any] = """\ """ __lowerCamelCase : Dict = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ __lowerCamelCase : Union[str, Any] = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def __UpperCamelCase( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "input_texts": datasets.Value("string" ), } ) , reference_urls=["https://huggingface.co/docs/transformers/perplexity"] , ) def __UpperCamelCase( self , A_ , A_ , A_ = 16 , A_ = True , A_=None ): '''simple docstring''' if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCamelCase : List[str] = "cuda" else: UpperCamelCase : Tuple = "cuda" if torch.cuda.is_available() else "cpu" UpperCamelCase : Any = AutoModelForCausalLM.from_pretrained(A_ ) UpperCamelCase : Any = model.to(A_ ) UpperCamelCase : Dict = AutoTokenizer.from_pretrained(A_ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCamelCase : Optional[Any] = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(A_ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"pad_token": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCamelCase : Dict = model.config.max_length - 1 else: UpperCamelCase : Union[str, Any] = model.config.max_length UpperCamelCase : Union[str, Any] = tokenizer( A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , return_tensors="pt" , return_attention_mask=A_ , ).to(A_ ) UpperCamelCase : int = encodings["input_ids"] UpperCamelCase : Any = encodings["attention_mask"] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCamelCase : Tuple = [] UpperCamelCase : Dict = CrossEntropyLoss(reduction="none" ) for start_index in logging.tqdm(range(0 , len(A_ ) , A_ ) ): UpperCamelCase : Any = min(start_index + batch_size , len(A_ ) ) UpperCamelCase : str = encoded_texts[start_index:end_index] UpperCamelCase : Tuple = attn_masks[start_index:end_index] if add_start_token: UpperCamelCase : Union[str, Any] = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(A_ ) UpperCamelCase : Any = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCamelCase : int = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(A_ ), attn_mask] , dim=1 ) UpperCamelCase : Dict = encoded_batch with torch.no_grad(): UpperCamelCase : Union[str, Any] = model(A_ , attention_mask=A_ ).logits UpperCamelCase : str = out_logits[..., :-1, :].contiguous() UpperCamelCase : Tuple = labels[..., 1:].contiguous() UpperCamelCase : Optional[int] = attn_mask[..., 1:].contiguous() UpperCamelCase : List[Any] = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , A_ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(A_ )}
38
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
1
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version __lowerCamelCase : Union[str, Any] = get_logger(__name__) class A__ : _UpperCAmelCase :Union[str, Any] = 'dummy_data' _UpperCAmelCase :Optional[int] = 'datasets' _UpperCAmelCase :Dict = False def __init__( self , A_ , A_ , A_ , A_ = None , A_ = False , A_ = True , A_ = None , ): '''simple docstring''' UpperCamelCase : List[str] = 0 UpperCamelCase : List[Any] = dataset_name UpperCamelCase : List[str] = cache_dir UpperCamelCase : Optional[int] = use_local_dummy_data UpperCamelCase : Union[str, Any] = config # download_callbacks take a single url as input UpperCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root UpperCamelCase : Optional[int] = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general UpperCamelCase : Optional[Any] = str(A_ ) # to be downloaded UpperCamelCase : List[Any] = None UpperCamelCase : Tuple = None @property def __UpperCamelCase( self ): '''simple docstring''' if self._dummy_file is None: UpperCamelCase : int = self.download_dummy_data() return self._dummy_file @property def __UpperCamelCase( self ): '''simple docstring''' if self.config is not None: # structure is dummy / config_name / version_name return os.path.join("dummy" , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join("dummy" , self.version_name ) @property def __UpperCamelCase( self ): '''simple docstring''' return os.path.join(self.dummy_data_folder , "dummy_data.zip" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) UpperCamelCase : List[Any] = cached_path( A_ , cache_dir=self.cache_dir , extract_compressed_file=A_ , force_extract=A_ ) return os.path.join(A_ , self.dummy_file_name ) @property def __UpperCamelCase( self ): '''simple docstring''' return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def __UpperCamelCase( self ): '''simple docstring''' if self._bucket_url is None: UpperCamelCase : List[str] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , "/" ) ) return self._bucket_url @property def __UpperCamelCase( self ): '''simple docstring''' if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , "/" ).split("/" )[:-1] ) def __UpperCamelCase( self , A_ , *A_ ): '''simple docstring''' if self.load_existing_dummy_data: # dummy data is downloaded and tested UpperCamelCase : Dict = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned UpperCamelCase : List[str] = self.dummy_file_name # special case when data_url is a dict if isinstance(A_ , A_ ): return self.create_dummy_data_dict(A_ , A_ ) elif isinstance(A_ , (list, tuple) ): return self.create_dummy_data_list(A_ , A_ ) else: return self.create_dummy_data_single(A_ , A_ ) def __UpperCamelCase( self , A_ , *A_ ): '''simple docstring''' return self.download_and_extract(A_ ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' return self.download_and_extract(A_ ) def __UpperCamelCase( self , A_ , *A_ , **A_ ): '''simple docstring''' return path def __UpperCamelCase( self ): '''simple docstring''' return {} def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(A_ , A_ ): for single_url in single_urls: download_callback(A_ ) else: UpperCamelCase : int = single_urls download_callback(A_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(A_ , A_ ): UpperCamelCase : Optional[Any] = [os.path.join(A_ , urllib.parse.quote_plus(Path(A_ ).name ) ) for x in single_urls] else: UpperCamelCase : List[str] = single_urls UpperCamelCase : int = os.path.join(A_ , urllib.parse.quote_plus(Path(A_ ).name ) ) UpperCamelCase : Optional[int] = value # make sure that values are unique if all(isinstance(A_ , A_ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique UpperCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Any = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one UpperCamelCase : str = all(bool(re.findall("[0-9]{3,}-of-[0-9]{3,}" , A_ ) ) for url in data_url ) UpperCamelCase : Union[str, Any] = all( url.startswith("https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed" ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): UpperCamelCase : List[Any] = [data_url[0]] * len(A_ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(A_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus UpperCamelCase : int = os.path.join(A_ , urllib.parse.quote_plus(single_url.split("/" )[-1] ) ) dummy_data_list.append(A_ ) return dummy_data_list def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' for download_callback in self.download_callbacks: download_callback(A_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus UpperCamelCase : List[str] = os.path.join(A_ , urllib.parse.quote_plus(data_url.split("/" )[-1] ) ) if os.path.exists(A_ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self , A_ ): '''simple docstring''' def _iter_archive_members(A_ ): # this preserves the order of the members inside the ZIP archive UpperCamelCase : List[str] = Path(self.dummy_file ).parent UpperCamelCase : Optional[Any] = path.relative_to(A_ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: UpperCamelCase : Union[str, Any] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(A_ ) UpperCamelCase : Tuple = Path(A_ ) UpperCamelCase : Any = _iter_archive_members(A_ ) if self.use_local_dummy_data else path.rglob("*" ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith((".", "__") ): yield file_path.relative_to(A_ ).as_posix(), file_path.open("rb" ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if not isinstance(A_ , A_ ): UpperCamelCase : Optional[int] = [paths] for path in paths: if os.path.isfile(A_ ): if os.path.basename(A_ ).startswith((".", "__") ): return yield path else: for dirpath, dirnames, filenames in os.walk(A_ ): if os.path.basename(A_ ).startswith((".", "__") ): continue dirnames.sort() for filename in sorted(A_ ): if filename.startswith((".", "__") ): continue yield os.path.join(A_ , A_ )
38
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
1
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand __lowerCamelCase : Optional[int] = logging.get_logger(__name__) # pylint: disable=invalid-name def A_ ( _lowerCAmelCase ) -> Any: if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(_lowerCAmelCase ): return ext raise Exception( F"""Unable to determine file format from file extension {path}. """ F"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Optional[int] = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) UpperCamelCase : Optional[int] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format UpperCamelCase : int = PipelineDataFormat.from_str( format=_lowerCAmelCase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(_lowerCAmelCase , _lowerCAmelCase ) class A__ ( __snake_case ): def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = nlp UpperCamelCase : Tuple = reader @staticmethod def __UpperCamelCase( A_ ): '''simple docstring''' UpperCamelCase : List[Any] = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=A_ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=A_ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=A_ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=A_ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=A_ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=A_ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=A_ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=A_ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : List[Any] = self._nlp, [] for entry in self._reader: UpperCamelCase : Tuple = nlp(**A_ ) if self._reader.is_multi_columns else nlp(A_ ) if isinstance(A_ , A_ ): outputs.append(A_ ) else: outputs += output # Saving data if self._nlp.binary_output: UpperCamelCase : Optional[Any] = self._reader.save_binary(A_ ) logger.warning(F"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(A_ )
38
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A_ ( ) -> Dict: UpperCamelCase : Tuple = ArgumentParser( description=( "PyTorch TPU distributed training launch " "helper utility that will spawn up " "multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCAmelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCAmelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCAmelCase ) return parser.parse_args() def A_ ( ) -> Optional[int]: UpperCamelCase : Tuple = parse_args() # Import training_script as a module. UpperCamelCase : Union[str, Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase : List[Any] = script_fpath.stem UpperCamelCase : Optional[Any] = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv UpperCamelCase : List[Any] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
38
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : Optional[Any] = { """configuration_bert""": ["""BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BertConfig""", """BertOnnxConfig"""], """tokenization_bert""": ["""BasicTokenizer""", """BertTokenizer""", """WordpieceTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Tuple = ["""BertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = [ """BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BertForMaskedLM""", """BertForMultipleChoice""", """BertForNextSentencePrediction""", """BertForPreTraining""", """BertForQuestionAnswering""", """BertForSequenceClassification""", """BertForTokenClassification""", """BertLayer""", """BertLMHeadModel""", """BertModel""", """BertPreTrainedModel""", """load_tf_weights_in_bert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Tuple = [ """TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBertEmbeddings""", """TFBertForMaskedLM""", """TFBertForMultipleChoice""", """TFBertForNextSentencePrediction""", """TFBertForPreTraining""", """TFBertForQuestionAnswering""", """TFBertForSequenceClassification""", """TFBertForTokenClassification""", """TFBertLMHeadModel""", """TFBertMainLayer""", """TFBertModel""", """TFBertPreTrainedModel""", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[int] = ["""TFBertTokenizer"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """FlaxBertForCausalLM""", """FlaxBertForMaskedLM""", """FlaxBertForMultipleChoice""", """FlaxBertForNextSentencePrediction""", """FlaxBertForPreTraining""", """FlaxBertForQuestionAnswering""", """FlaxBertForSequenceClassification""", """FlaxBertForTokenClassification""", """FlaxBertModel""", """FlaxBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys __lowerCamelCase : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
def A_ ( _lowerCAmelCase ) -> int: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("only integers accepted as input" ) else: UpperCamelCase : Any = str(abs(_lowerCAmelCase ) ) UpperCamelCase : Dict = [list(_lowerCAmelCase ) for char in range(len(_lowerCAmelCase ) )] for index in range(len(_lowerCAmelCase ) ): num_transpositions[index].pop(_lowerCAmelCase ) return max( int("".join(list(_lowerCAmelCase ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__("""doctest""").testmod()
38
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
1
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str: if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) UpperCamelCase : Any = str(bin(_lowerCAmelCase ) )[2:] # remove the leading "0b" UpperCamelCase : List[str] = str(bin(_lowerCAmelCase ) )[2:] # remove the leading "0b" UpperCamelCase : List[str] = max(len(_lowerCAmelCase ) , len(_lowerCAmelCase ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(_lowerCAmelCase ) , b_binary.zfill(_lowerCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __lowerCamelCase : Dict = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __lowerCamelCase : Tuple = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = VOCAB_FILES_NAMES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[Any] = BlenderbotSmallTokenizer def __init__( self , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , A_=True , **A_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=A_ , merges=A_ , add_prefix_space=A_ , trim_offsets=A_ , ) , bos_token=A_ , eos_token=A_ , unk_token=A_ , **A_ , ) UpperCamelCase : Union[str, Any] = add_prefix_space def __UpperCamelCase( self , A_ , A_=None ): '''simple docstring''' UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
38
1
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : int = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
import math import random from typing import Any from .hill_climbing import SearchProblem def A_ ( _lowerCAmelCase , _lowerCAmelCase = True , _lowerCAmelCase = math.inf , _lowerCAmelCase = -math.inf , _lowerCAmelCase = math.inf , _lowerCAmelCase = -math.inf , _lowerCAmelCase = False , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.01 , _lowerCAmelCase = 1 , ) -> Any: UpperCamelCase : int = False UpperCamelCase : Tuple = search_prob UpperCamelCase : str = start_temperate UpperCamelCase : Any = [] UpperCamelCase : str = 0 UpperCamelCase : str = None while not search_end: UpperCamelCase : Any = current_state.score() if best_state is None or current_score > best_state.score(): UpperCamelCase : List[Any] = current_state scores.append(_lowerCAmelCase ) iterations += 1 UpperCamelCase : List[str] = None UpperCamelCase : Dict = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to UpperCamelCase : List[Any] = random.randint(0 , len(_lowerCAmelCase ) - 1 ) # picking a random neighbor UpperCamelCase : Dict = neighbors.pop(_lowerCAmelCase ) UpperCamelCase : Any = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: UpperCamelCase : int = change * -1 # in case we are finding minimum if change > 0: # improves the solution UpperCamelCase : List[str] = picked_neighbor else: UpperCamelCase : List[str] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability UpperCamelCase : Optional[Any] = picked_neighbor UpperCamelCase : Any = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor UpperCamelCase : Dict = True else: UpperCamelCase : List[str] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(_lowerCAmelCase ) , _lowerCAmelCase ) plt.xlabel("Iterations" ) plt.ylabel("Function values" ) plt.show() return best_state if __name__ == "__main__": def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) __lowerCamelCase : Optional[int] = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __lowerCamelCase : Union[str, Any] = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( """The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ f"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) __lowerCamelCase : List[Any] = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) __lowerCamelCase : List[Any] = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( """The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ f"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Any: return (3 * x**2) - (6 * y) __lowerCamelCase : int = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __lowerCamelCase : int = simulated_annealing(prob, find_max=False, visualization=True) print( """The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ f"""{local_min.score()}""" ) __lowerCamelCase : Tuple = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) __lowerCamelCase : int = simulated_annealing(prob, find_max=True, visualization=True) print( """The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ f"""{local_min.score()}""" )
38
import logging import os import threading import time try: import warnings except ImportError: __lowerCamelCase : str = None try: import msvcrt except ImportError: __lowerCamelCase : str = None try: import fcntl except ImportError: __lowerCamelCase : List[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __lowerCamelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ __lowerCamelCase : str = [ """Timeout""", """BaseFileLock""", """WindowsFileLock""", """UnixFileLock""", """SoftFileLock""", """FileLock""", ] __lowerCamelCase : Union[str, Any] = """3.0.12""" __lowerCamelCase : Any = None def A_ ( ) -> List[Any]: global _logger UpperCamelCase : Any = _logger or logging.getLogger(__name__ ) return _logger class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = lock_file return None def __str__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = lock return None def __enter__( self ): '''simple docstring''' return self.lock def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.lock.release() return None class A__ : def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : List[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long UpperCamelCase : Dict = self.hash_filename_if_too_long(A_ , A_ ) # The path to the lock file. UpperCamelCase : List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. UpperCamelCase : Tuple = None # The default timeout value. UpperCamelCase : Optional[Any] = timeout # We use this lock primarily for the lock counter. UpperCamelCase : Union[str, Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. UpperCamelCase : Dict = 0 return None @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file @property def __UpperCamelCase( self ): '''simple docstring''' return self._timeout @timeout.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = float(A_ ) return None def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file_fd is not None def __UpperCamelCase( self , A_=None , A_=0.05 ): '''simple docstring''' if timeout is None: UpperCamelCase : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 UpperCamelCase : Dict = id(self ) UpperCamelCase : List[str] = self._lock_file UpperCamelCase : int = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(A_ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: UpperCamelCase : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __UpperCamelCase( self , A_=False ): '''simple docstring''' with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: UpperCamelCase : List[Any] = id(self ) UpperCamelCase : Dict = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() UpperCamelCase : Dict = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): '''simple docstring''' self.acquire() return self def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.release() return None def __del__( self ): '''simple docstring''' self.release(force=A_ ) return None def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = os.path.basename(A_ ) if len(A_ ) > max_length and max_length > 0: UpperCamelCase : Optional[int] = os.path.dirname(A_ ) UpperCamelCase : int = str(hash(A_ ) ) UpperCamelCase : Any = filename[: max_length - len(A_ ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(A_ , A_ ) else: return path class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' from .file_utils import relative_to_absolute_path super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) UpperCamelCase : List[Any] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: UpperCamelCase : str = os.open(self._lock_file , A_ ) except OSError: pass else: try: msvcrt.locking(A_ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(A_ ) else: UpperCamelCase : Optional[Any] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self._lock_file_fd UpperCamelCase : str = None msvcrt.locking(A_ , msvcrt.LK_UNLCK , 1 ) os.close(A_ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : Tuple = os.statvfs(os.path.dirname(A_ ) ).f_namemax super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC UpperCamelCase : int = os.open(self._lock_file , A_ ) try: fcntl.flock(A_ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(A_ ) else: UpperCamelCase : List[str] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self._lock_file_fd UpperCamelCase : List[Any] = None fcntl.flock(A_ , fcntl.LOCK_UN ) os.close(A_ ) return None class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: UpperCamelCase : Optional[int] = os.open(self._lock_file , A_ ) except OSError: pass else: UpperCamelCase : Tuple = fd return None def __UpperCamelCase( self ): '''simple docstring''' os.close(self._lock_file_fd ) UpperCamelCase : str = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __lowerCamelCase : Dict = None if msvcrt: __lowerCamelCase : Any = WindowsFileLock elif fcntl: __lowerCamelCase : Any = UnixFileLock else: __lowerCamelCase : int = SoftFileLock if warnings is not None: warnings.warn("""only soft file lock is available""")
38
1
import requests def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> None: UpperCamelCase : List[Any] = {"Content-Type": "application/json"} UpperCamelCase : Dict = requests.post(_lowerCAmelCase , json={"text": message_body} , headers=_lowerCAmelCase ) if response.status_code != 200: UpperCamelCase : Tuple = ( "Request to slack returned an error " F"""{response.status_code}, the response is:\n{response.text}""" ) raise ValueError(_lowerCAmelCase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message("""<YOUR MESSAGE BODY>""", """<SLACK CHANNEL URL>""")
38
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> str: if config_name_or_path is None: UpperCamelCase : Dict = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCamelCase : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCamelCase : Tuple = question_encoder_name_or_path UpperCamelCase : Any = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCamelCase : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Tuple = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : int = gen_config UpperCamelCase : Dict = question_encoder_config UpperCamelCase : Tuple = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--model_type""", choices=["""rag_sequence""", """rag_token"""], required=True, type=str, help="""RAG model type: rag_sequence, rag_token""", ) parser.add_argument("""--dest""", type=str, required=True, help="""Path to the output checkpoint directory.""") parser.add_argument("""--generator_name_or_path""", type=str, required=True, help="""Generator model identifier""") parser.add_argument( """--question_encoder_name_or_path""", type=str, required=True, help="""Question encoder model identifier""" ) parser.add_argument( """--generator_tokenizer_name_or_path""", type=str, help="""Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``""", ) parser.add_argument( """--question_encoder_tokenizer_name_or_path""", type=str, help="""Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``""", ) parser.add_argument( """--config_name_or_path""", type=str, help=( """Identifier of the model config to use, if not provided, resolves to a base config for a given""" """ ``model_type``""" ), ) __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
38
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Any = logging.get_logger(__name__) def A_ ( _lowerCAmelCase , _lowerCAmelCase=False ) -> Union[str, Any]: UpperCamelCase : Any = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight""") ) rename_keys.append((F"""patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias""", F"""vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias""") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" UpperCamelCase : str = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False ) -> Dict: for i in range(config.num_hidden_layers ): if base_model: UpperCamelCase : int = "" else: UpperCamelCase : Dict = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCamelCase : int = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) UpperCamelCase : Any = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCamelCase : str = in_proj_weight[ : config.hidden_size, : ] UpperCamelCase : int = in_proj_bias[: config.hidden_size] UpperCamelCase : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCamelCase : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCamelCase : Tuple = in_proj_weight[ -config.hidden_size :, : ] UpperCamelCase : Any = in_proj_bias[-config.hidden_size :] def A_ ( _lowerCAmelCase ) -> Tuple: UpperCamelCase : str = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Any: UpperCamelCase : Any = dct.pop(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = val def A_ ( ) -> Any: UpperCamelCase : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCamelCase : str = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False ) -> List[str]: UpperCamelCase : Any = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_lowerCAmelCase , ) UpperCamelCase : List[Any] = ViTHybridConfig(backbone_config=_lowerCAmelCase , image_size=384 , num_labels=1000 ) UpperCamelCase : Tuple = False # load original model from timm UpperCamelCase : int = timm.create_model(_lowerCAmelCase , pretrained=_lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCamelCase : Optional[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_lowerCAmelCase ) UpperCamelCase : Any = create_rename_keys(_lowerCAmelCase , _lowerCAmelCase ) for src, dest in rename_keys: rename_key(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : Union[str, Any] = "huggingface/label-files" UpperCamelCase : List[str] = "imagenet-1k-id2label.json" UpperCamelCase : int = json.load(open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="dataset" ) , "r" ) ) UpperCamelCase : List[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCamelCase : Tuple = idalabel UpperCamelCase : Optional[Any] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": UpperCamelCase : Optional[int] = ViTHybridModel(_lowerCAmelCase ).eval() else: UpperCamelCase : Any = ViTHybridForImageClassification(_lowerCAmelCase ).eval() model.load_state_dict(_lowerCAmelCase ) # create image processor UpperCamelCase : Dict = create_transform(**resolve_data_config({} , model=_lowerCAmelCase ) ) UpperCamelCase : Union[str, Any] = transform.transforms UpperCamelCase : List[Any] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } UpperCamelCase : Tuple = ViTHybridImageProcessor( do_resize=_lowerCAmelCase , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_lowerCAmelCase , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_lowerCAmelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) UpperCamelCase : Tuple = prepare_img() UpperCamelCase : Dict = transform(_lowerCAmelCase ).unsqueeze(0 ) UpperCamelCase : Tuple = processor(_lowerCAmelCase , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_lowerCAmelCase , _lowerCAmelCase ) # verify logits with torch.no_grad(): UpperCamelCase : Optional[Any] = model(_lowerCAmelCase ) UpperCamelCase : str = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: UpperCamelCase : Union[str, Any] = timm_model.forward_features(_lowerCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_lowerCAmelCase , outputs.pooler_output , atol=1e-3 ) else: UpperCamelCase : List[str] = timm_model(_lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCAmelCase , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: print(F"""Pushing model and processor to the hub {vit_name}""" ) model.push_to_hub(F"""ybelkada/{vit_name}""" ) processor.push_to_hub(F"""ybelkada/{vit_name}""" ) if __name__ == "__main__": __lowerCamelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_r50_s16_384""", type=str, help="""Name of the hybrid ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) __lowerCamelCase : Union[str, Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
38
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : str = 13 UpperCamelCase : int = 7 UpperCamelCase : str = True UpperCamelCase : Dict = True UpperCamelCase : str = True UpperCamelCase : Tuple = True UpperCamelCase : List[str] = 99 UpperCamelCase : Optional[Any] = 384 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Dict = 37 UpperCamelCase : Any = "gelu" UpperCamelCase : List[Any] = 0.1 UpperCamelCase : int = 0.1 UpperCamelCase : Tuple = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : int = 2 UpperCamelCase : Dict = 0.02 UpperCamelCase : Optional[Any] = 3 UpperCamelCase : List[Any] = 4 UpperCamelCase : Dict = 128 UpperCamelCase : Optional[Any] = 2 UpperCamelCase : Optional[int] = 9 UpperCamelCase : Optional[int] = 1 UpperCamelCase : Union[str, Any] = None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None if self.use_token_type_ids: UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel(config=A_ ) UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : Any = model(A_ ) UpperCamelCase : int = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = TFConvBertForMaskedLM(config=A_ ) UpperCamelCase : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = TFConvBertForSequenceClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : str = TFConvBertForMultipleChoice(config=A_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Dict = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Any = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : str = TFConvBertForTokenClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : str = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFConvBertForQuestionAnswering(config=A_ ) UpperCamelCase : Union[str, Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Union[str, Any] = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase :Any = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = TFConvBertModelTester(self ) UpperCamelCase : Dict = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Optional[Any] = True UpperCamelCase : Any = True if hasattr(A_ , "use_cache" ): UpperCamelCase : List[str] = True UpperCamelCase : List[Any] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Any = getattr(self.model_tester , "key_length" , A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[Any] = self._prepare_for_class(A_ , A_ ) UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Optional[int] = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) UpperCamelCase : Union[str, Any] = os.path.join(A_ , "saved_model" , "1" ) UpperCamelCase : Dict = tf.keras.models.load_model(A_ ) UpperCamelCase : str = model(A_ ) if self.is_encoder_decoder: UpperCamelCase : Union[str, Any] = outputs["encoder_hidden_states"] UpperCamelCase : Any = outputs["encoder_attentions"] else: UpperCamelCase : Any = outputs["hidden_states"] UpperCamelCase : List[str] = outputs["attentions"] self.assertEqual(len(A_ ) , A_ ) UpperCamelCase : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True UpperCamelCase : int = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "key_length" , A_ ) UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , A_ ) def check_decoder_attentions_output(A_ ): UpperCamelCase : Optional[Any] = len(A_ ) self.assertEqual(out_len % 2 , 0 ) UpperCamelCase : Any = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): UpperCamelCase : Dict = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True UpperCamelCase : List[Any] = False UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : List[str] = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Tuple = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase : Tuple = True UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine UpperCamelCase : Optional[int] = True UpperCamelCase : List[str] = True UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : Optional[Any] = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : List[str] = model(A_ )[0] UpperCamelCase : int = [1, 6, 768] self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
38
1
import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): __lowerCamelCase : int = yaml.safe_load( """\ name: \"\" allow_empty: false allow_empty_text: true subsections: - name: \"Dataset Card for X\" # First-level markdown heading allow_empty: false allow_empty_text: true subsections: - name: \"Table of Contents\" allow_empty: false allow_empty_text: false subsections: null - name: \"Dataset Description\" allow_empty: false allow_empty_text: false subsections: - name: \"Dataset Summary\" allow_empty: false allow_empty_text: false subsections: null - name: \"Supported Tasks and Leaderboards\" allow_empty: true allow_empty_text: true subsections: null - name: Languages allow_empty: false allow_empty_text: true subsections: null """ ) __lowerCamelCase : int = { """name""": """root""", """text""": """""", """is_empty_text""": True, """subsections""": [ { """name""": """Dataset Card for My Dataset""", """text""": """""", """is_empty_text""": True, """subsections""": [ {"""name""": """Table of Contents""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": []}, { """name""": """Dataset Description""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Dataset Summary""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [], }, { """name""": """Supported Tasks and Leaderboards""", """text""": """""", """is_empty_text""": True, """subsections""": [], }, {"""name""": """Languages""", """text""": """Language Text""", """is_empty_text""": False, """subsections""": []}, ], }, ], } ], } __lowerCamelCase : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : List[str] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. #### Extra Ignored Subsection ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : int = { """name""": """root""", """text""": """""", """is_empty_text""": True, """subsections""": [ { """name""": """Dataset Card for My Dataset""", """text""": """""", """is_empty_text""": True, """subsections""": [ {"""name""": """Table of Contents""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": []}, { """name""": """Dataset Description""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Dataset Summary""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Extra Ignored Subsection""", """text""": """""", """is_empty_text""": True, """subsections""": [], } ], }, { """name""": """Supported Tasks and Leaderboards""", """text""": """""", """is_empty_text""": True, """subsections""": [], }, {"""name""": """Languages""", """text""": """Language Text""", """is_empty_text""": False, """subsections""": []}, ], }, ], } ], } __lowerCamelCase : int = """\ --- --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : str = ( """The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.""" ) __lowerCamelCase : List[Any] = """\ # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Dict = ( """The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.""" ) __lowerCamelCase : Optional[Any] = """\ --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Optional[Any] = """The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.""" __lowerCamelCase : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Dict = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).""" __lowerCamelCase : Tuple = """\ --- language: - zh - en --- # Dataset Card for My Dataset """ __lowerCamelCase : List[str] = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found 'None'.""" __lowerCamelCase : List[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Languages Language Text """ __lowerCamelCase : str = """The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.""" __lowerCamelCase : List[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages """ __lowerCamelCase : Dict = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.""" __lowerCamelCase : Optional[int] = """\ --- language: - zh - en --- ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Tuple = """The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.""" __lowerCamelCase : Optional[int] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text # Dataset Card My Dataset """ __lowerCamelCase : Optional[Any] = """The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.""" __lowerCamelCase : Optional[int] = """\ --- language: - zh - en --- # Dataset Card My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Union[str, Any] = """The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.""" __lowerCamelCase : Union[str, Any] = """""" __lowerCamelCase : List[str] = """The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.""" __lowerCamelCase : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ __lowerCamelCase : Optional[Any] = """The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.""" @pytest.mark.parametrize( "readme_md, expected_dict" , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: assert ReadMe.from_string(_lowerCAmelCase , _lowerCAmelCase ).to_dict() == expected_dict @pytest.mark.parametrize( "readme_md, expected_error" , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: with pytest.raises(_lowerCAmelCase , match=re.escape(expected_error.format(path="root" ) ) ): UpperCamelCase : List[str] = ReadMe.from_string(_lowerCAmelCase , _lowerCAmelCase ) readme.validate() @pytest.mark.parametrize( "readme_md, expected_error" , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[int]: with pytest.raises(_lowerCAmelCase , match=re.escape(expected_error.format(path="root" ) ) ): ReadMe.from_string(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( "readme_md," , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def A_ ( _lowerCAmelCase ) -> Optional[int]: ReadMe.from_string(_lowerCAmelCase , _lowerCAmelCase , suppress_parsing_errors=_lowerCAmelCase ) @pytest.mark.parametrize( "readme_md, expected_dict" , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: with tempfile.TemporaryDirectory() as tmp_dir: UpperCamelCase : str = Path(_lowerCAmelCase ) / "README.md" with open(_lowerCAmelCase , "w+" ) as readme_file: readme_file.write(_lowerCAmelCase ) UpperCamelCase : Any = ReadMe.from_readme(_lowerCAmelCase , _lowerCAmelCase ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( "readme_md, expected_error" , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Dict: with tempfile.TemporaryDirectory() as tmp_dir: UpperCamelCase : Dict = Path(_lowerCAmelCase ) / "README.md" with open(_lowerCAmelCase , "w+" ) as readme_file: readme_file.write(_lowerCAmelCase ) UpperCamelCase : Optional[int] = expected_error.format(path=_lowerCAmelCase ) with pytest.raises(_lowerCAmelCase , match=re.escape(_lowerCAmelCase ) ): UpperCamelCase : str = ReadMe.from_readme(_lowerCAmelCase , _lowerCAmelCase ) readme.validate() @pytest.mark.parametrize( "readme_md, expected_error" , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: with tempfile.TemporaryDirectory() as tmp_dir: UpperCamelCase : Union[str, Any] = Path(_lowerCAmelCase ) / "README.md" with open(_lowerCAmelCase , "w+" ) as readme_file: readme_file.write(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = expected_error.format(path=_lowerCAmelCase ) with pytest.raises(_lowerCAmelCase , match=re.escape(_lowerCAmelCase ) ): ReadMe.from_readme(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( "readme_md," , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def A_ ( _lowerCAmelCase ) -> Optional[int]: with tempfile.TemporaryDirectory() as tmp_dir: UpperCamelCase : int = Path(_lowerCAmelCase ) / "README.md" with open(_lowerCAmelCase , "w+" ) as readme_file: readme_file.write(_lowerCAmelCase ) ReadMe.from_readme(_lowerCAmelCase , _lowerCAmelCase , suppress_parsing_errors=_lowerCAmelCase )
38
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : str = { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'camembert' def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : List[str] = vocab_size UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : Dict = hidden_act UpperCamelCase : str = intermediate_size UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : int = initializer_range UpperCamelCase : List[str] = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : int = use_cache UpperCamelCase : List[str] = classifier_dropout class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : str = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
38
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: return int(input_a == input_a == 0 ) def A_ ( ) -> None: print("Truth Table of NOR Gate:" ) print("| Input 1 | Input 2 | Output |" ) print(F"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Dict: print("\nThe shortest path matrix using Floyd Warshall algorithm\n" ) for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): if dist[i][j] != float("inf" ): print(int(dist[i][j] ) , end="\t" ) else: print("INF" , end="\t" ) print() def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: UpperCamelCase : Dict = [[float("inf" ) for _ in range(_lowerCAmelCase )] for _ in range(_lowerCAmelCase )] for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(_lowerCAmelCase ): # looping through rows of graph array for i in range(_lowerCAmelCase ): # looping through columns of graph array for j in range(_lowerCAmelCase ): if ( dist[i][k] != float("inf" ) and dist[k][j] != float("inf" ) and dist[i][k] + dist[k][j] < dist[i][j] ): UpperCamelCase : Union[str, Any] = dist[i][k] + dist[k][j] _print_dist(_lowerCAmelCase , _lowerCAmelCase ) return dist, v if __name__ == "__main__": __lowerCamelCase : Optional[Any] = int(input("""Enter number of vertices: """)) __lowerCamelCase : int = int(input("""Enter number of edges: """)) __lowerCamelCase : Any = [[float("""inf""") for i in range(v)] for j in range(v)] for i in range(v): __lowerCamelCase : Dict = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print("""\nEdge """, i + 1) __lowerCamelCase : Union[str, Any] = int(input("""Enter source:""")) __lowerCamelCase : Dict = int(input("""Enter destination:""")) __lowerCamelCase : str = float(input("""Enter weight:""")) __lowerCamelCase : str = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
38
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __snake_case ): _UpperCAmelCase :Optional[int] = ['image_processor', 'tokenizer'] _UpperCAmelCase :Tuple = 'BlipImageProcessor' _UpperCAmelCase :Optional[int] = 'AutoTokenizer' def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = False super().__init__(A_ , A_ ) UpperCamelCase : str = self.image_processor def __call__( self , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: UpperCamelCase : int = self.tokenizer UpperCamelCase : Optional[int] = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) return text_encoding # add pixel_values UpperCamelCase : int = self.image_processor(A_ , return_tensors=A_ ) if text is not None: UpperCamelCase : Dict = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) else: UpperCamelCase : Dict = None if text_encoding is not None: encoding_image_processor.update(A_ ) return encoding_image_processor def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.batch_decode(*A_ , **A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.decode(*A_ , **A_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.tokenizer.model_input_names UpperCamelCase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
38
1
from math import factorial, pi def A_ ( _lowerCAmelCase , _lowerCAmelCase = 30 ) -> float: if not isinstance(_lowerCAmelCase , (int, float) ): raise ValueError("maclaurin_sin() requires either an int or float for theta" ) if not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or accuracy <= 0: raise ValueError("maclaurin_sin() requires a positive int for accuracy" ) UpperCamelCase : str = float(_lowerCAmelCase ) UpperCamelCase : Optional[Any] = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase , _lowerCAmelCase = 30 ) -> float: if not isinstance(_lowerCAmelCase , (int, float) ): raise ValueError("maclaurin_cos() requires either an int or float for theta" ) if not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or accuracy <= 0: raise ValueError("maclaurin_cos() requires a positive int for accuracy" ) UpperCamelCase : Dict = float(_lowerCAmelCase ) UpperCamelCase : int = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(_lowerCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(10)) print(maclaurin_sin(-10)) print(maclaurin_sin(10, 15)) print(maclaurin_sin(-10, 15)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(10, 15)) print(maclaurin_cos(-10, 15))
38
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __lowerCamelCase : Dict = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Tuple = ['audio_values', 'audio_mask'] def __init__( self , A_=2048 , A_=1 , A_=[16, 16] , A_=128 , A_=4_4100 , A_=86 , A_=2048 , A_=0.0 , **A_ , ): '''simple docstring''' super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , **A_ , ) UpperCamelCase : Optional[int] = spectrogram_length UpperCamelCase : Dict = num_channels UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : str = feature_size // self.patch_size[1] UpperCamelCase : List[str] = n_fft UpperCamelCase : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase : Optional[int] = sampling_rate UpperCamelCase : int = padding_value UpperCamelCase : str = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=A_ , norm="slaney" , mel_scale="slaney" , ).T def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = spectrogram( A_ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) UpperCamelCase : List[Any] = log_spec[:, :-1] UpperCamelCase : Optional[int] = log_spec - 20.0 UpperCamelCase : str = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A_ , A_ = None , A_ = True , A_ = None , A_ = False , A_ = False , **A_ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase : Optional[int] = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase : Union[str, Any] = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): UpperCamelCase : str = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase : str = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A_ ): UpperCamelCase : int = [np.asarray(A_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase : Tuple = np.array(A_ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase : Union[str, Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase : Any = np.ones([len(A_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase : List[str] = padded_audio_features * self.padding_value for i in range(len(A_ ) ): UpperCamelCase : Union[str, Any] = audio_features[i] UpperCamelCase : Optional[int] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase : Optional[Any] = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: UpperCamelCase : int = {"audio_values": padded_audio_features} UpperCamelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
38
1
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __snake_case ): _UpperCAmelCase :Optional[int] = ['image_processor', 'tokenizer'] _UpperCAmelCase :Tuple = 'BlipImageProcessor' _UpperCAmelCase :Optional[int] = 'AutoTokenizer' def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = False super().__init__(A_ , A_ ) UpperCamelCase : str = self.image_processor def __call__( self , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: UpperCamelCase : int = self.tokenizer UpperCamelCase : Optional[int] = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) return text_encoding # add pixel_values UpperCamelCase : int = self.image_processor(A_ , return_tensors=A_ ) if text is not None: UpperCamelCase : Dict = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) else: UpperCamelCase : Dict = None if text_encoding is not None: encoding_image_processor.update(A_ ) return encoding_image_processor def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.batch_decode(*A_ , **A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.decode(*A_ , **A_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.tokenizer.model_input_names UpperCamelCase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
38
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Dict = TypeVar("""KT""") __lowerCamelCase : Dict = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self , A_ = "root" , A_ = None ): '''simple docstring''' UpperCamelCase : int = key UpperCamelCase : List[Any] = value UpperCamelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F"""Node({self.key}: {self.value})""" @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self , A_ = 0.5 , A_ = 16 ): '''simple docstring''' UpperCamelCase : Node[KT, VT] = Node[KT, VT]() UpperCamelCase : List[Any] = 0 UpperCamelCase : Union[str, Any] = p UpperCamelCase : List[str] = max_level def __str__( self ): '''simple docstring''' UpperCamelCase : int = list(self ) if len(A_ ) == 0: return F"""SkipList(level={self.level})""" UpperCamelCase : str = max((len(str(A_ ) ) for item in items) , default=4 ) UpperCamelCase : Dict = max(A_ , 4 ) + 4 UpperCamelCase : str = self.head UpperCamelCase : List[Any] = [] UpperCamelCase : int = node.forward.copy() lines.append(F"""[{node.key}]""".ljust(A_ , "-" ) + "* " * len(A_ ) ) lines.append(" " * label_size + "| " * len(A_ ) ) while len(node.forward ) != 0: UpperCamelCase : Union[str, Any] = node.forward[0] lines.append( F"""[{node.key}]""".ljust(A_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(A_ ) ) UpperCamelCase : Tuple = node.forward lines.append("None".ljust(A_ ) + "* " * len(A_ ) ) return F"""SkipList(level={self.level})\n""" + "\n".join(A_ ) def __iter__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCamelCase : Union[str, Any] = node.forward[0] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = 1 while random() < self.p and level < self.max_level: level += 1 return level def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = [] UpperCamelCase : List[Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCamelCase : str = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(A_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self._locate_node(A_ ) if node is not None: for i, update_node in enumerate(A_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCamelCase : Tuple = node.forward[i] else: UpperCamelCase : List[Any] = update_node.forward[:i] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self._locate_node(A_ ) if node is not None: UpperCamelCase : Union[str, Any] = value else: UpperCamelCase : Dict = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , A_ ): update_vector.append(self.head ) UpperCamelCase : Optional[int] = level UpperCamelCase : Dict = Node(A_ , A_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(A_ ) else: UpperCamelCase : List[Any] = new_node def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self._locate_node(A_ ) if node is not None: return node.value return None def A_ ( ) -> List[Any]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) UpperCamelCase : Optional[int] = skip_list.head UpperCamelCase : List[str] = {} while node.level != 0: UpperCamelCase : str = node.forward[0] UpperCamelCase : Optional[int] = node.value assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def A_ ( ) -> List[Any]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) UpperCamelCase : Dict = skip_list.head UpperCamelCase : Tuple = {} while node.level != 0: UpperCamelCase : List[str] = node.forward[0] UpperCamelCase : Dict = node.value if len(_lowerCAmelCase ) != 4: print() assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() assert skip_list.find("Some key" ) is None def A_ ( ) -> Tuple: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def A_ ( ) -> Dict: UpperCamelCase : Optional[int] = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def A_ ( ) -> Dict: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[str]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_lowerCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_lowerCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def A_ ( ) -> Union[str, Any]: def is_sorted(_lowerCAmelCase ): return all(next_item >= item for item, next_item in zip(_lowerCAmelCase , lst[1:] ) ) UpperCamelCase : int = SkipList() for i in range(10 ): skip_list.insert(_lowerCAmelCase , _lowerCAmelCase ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_lowerCAmelCase ) ) def A_ ( ) -> Tuple: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def A_ ( ) -> List[str]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu __lowerCamelCase : Union[str, Any] = [ """EAGER""", """AOT_EAGER""", """INDUCTOR""", """NVFUSER""", """AOT_NVFUSER""", """AOT_CUDAGRAPHS""", """OFI""", """FX2TRT""", """ONNXRT""", """IPEX""", ] def A_ ( _lowerCAmelCase , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None ) -> str: UpperCamelCase : int = True while ask_again: UpperCamelCase : Dict = input(_lowerCAmelCase ) try: if default is not None and len(_lowerCAmelCase ) == 0: return default return convert_value(_lowerCAmelCase ) if convert_value is not None else result except Exception: if error_message is not None: print(_lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=[] , _lowerCAmelCase=None , _lowerCAmelCase=0 ) -> List[Any]: UpperCamelCase : int = BulletMenu(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : Optional[int] = menu.run(default_choice=_lowerCAmelCase ) return convert_value(_lowerCAmelCase ) if convert_value is not None else result def A_ ( _lowerCAmelCase ) -> Optional[int]: UpperCamelCase : Any = int(_lowerCAmelCase ) return ComputeEnvironment(["LOCAL_MACHINE", "AMAZON_SAGEMAKER"][value] ) def A_ ( _lowerCAmelCase ) -> Union[str, Any]: UpperCamelCase : Dict = int(_lowerCAmelCase ) return DistributedType(["NO", "MULTI_CPU", "MULTI_XPU", "MULTI_GPU", "MULTI_NPU", "TPU"][value] ) def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : int = int(_lowerCAmelCase ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Dict = int(_lowerCAmelCase ) return PrecisionType(["no", "fp16", "bf16", "fp8"][value] ) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : List[Any] = int(_lowerCAmelCase ) return SageMakerDistributedType(["NO", "DATA_PARALLEL", "MODEL_PARALLEL"][value] ) def A_ ( _lowerCAmelCase ) -> Any: return {"yes": True, "no": False}[value.lower()] class A__ ( argparse.RawDescriptionHelpFormatter ): def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = super()._format_usage(A_ , A_ , A_ , A_ ) UpperCamelCase : List[Any] = usage.replace("<command> [<args>] " , "" ) return usage
38
from PIL import Image def A_ ( _lowerCAmelCase ) -> Image: UpperCamelCase , UpperCamelCase : List[Any] = image.size UpperCamelCase : Union[str, Any] = 0 UpperCamelCase : List[str] = image.load() for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(_lowerCAmelCase ): for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
38
1
def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> bool: return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> bool: # Base Case if index == len(_lowerCAmelCase ): return True # Recursive Step for i in range(_lowerCAmelCase ): if valid_coloring(graph[index] , _lowerCAmelCase , _lowerCAmelCase ): # Color current vertex UpperCamelCase : Optional[Any] = i # Validate coloring if util_color(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , index + 1 ): return True # Backtrack UpperCamelCase : int = -1 return False def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> list[int]: UpperCamelCase : int = [-1] * len(_lowerCAmelCase ) if util_color(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , 0 ): return colored_vertices return []
38
from math import loga def A_ ( _lowerCAmelCase ) -> int: if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
1
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class A__ ( __snake_case ): def __init__( self , A_ , A_ , A_=1024 , A_=1024 , A_=3.6 ): '''simple docstring''' UpperCamelCase : List[Any] = tokenizer UpperCamelCase : Optional[int] = tokenizer.bos_token_id UpperCamelCase : Optional[int] = dataset UpperCamelCase : Optional[int] = seq_length UpperCamelCase : List[str] = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' UpperCamelCase : List[str] = iter(self.dataset ) UpperCamelCase : Union[str, Any] = True while more_examples: UpperCamelCase , UpperCamelCase : Union[str, Any] = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(A_ )["content"] ) buffer_len += len(buffer[-1] ) except StopIteration: UpperCamelCase : Optional[int] = False break UpperCamelCase : Dict = tokenizer(A_ , truncation=A_ )["input_ids"] UpperCamelCase : List[str] = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(A_ ) , self.seq_length ): UpperCamelCase : Optional[Any] = all_token_ids[i : i + self.seq_length] if len(A_ ) == self.seq_length: yield torch.tensor(A_ ) def A_ ( _lowerCAmelCase ) -> Optional[int]: UpperCamelCase : Union[str, Any] = {"streaming": True} UpperCamelCase : Optional[Any] = load_dataset(args.dataset_name , split="train" , **_lowerCAmelCase ) UpperCamelCase : List[str] = ConstantLengthDataset(_lowerCAmelCase , _lowerCAmelCase , seq_length=args.seq_length ) UpperCamelCase : List[str] = DataLoader(_lowerCAmelCase , batch_size=args.batch_size ) return eval_dataloader def A_ ( _lowerCAmelCase ) -> Optional[Any]: model.eval() UpperCamelCase : Any = [] for step, batch in enumerate(_lowerCAmelCase ): with torch.no_grad(): UpperCamelCase : str = model(_lowerCAmelCase , labels=_lowerCAmelCase ) UpperCamelCase : Dict = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_lowerCAmelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break UpperCamelCase : Optional[int] = torch.mean(torch.cat(_lowerCAmelCase ) ) try: UpperCamelCase : int = torch.exp(_lowerCAmelCase ) except OverflowError: UpperCamelCase : Optional[Any] = float("inf" ) return loss.item(), perplexity.item() # Setup Accelerator __lowerCamelCase : Tuple = Accelerator() # Parse configuration __lowerCamelCase : Tuple = HfArgumentParser(EvaluationArguments) __lowerCamelCase : List[str] = parser.parse_args() set_seed(args.seed) # Logging __lowerCamelCase : Union[str, Any] = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer __lowerCamelCase : Optional[Any] = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __lowerCamelCase : Any = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __lowerCamelCase : Union[str, Any] = create_dataloader(args) # Prepare everything with our `accelerator`. __lowerCamelCase , __lowerCamelCase : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") __lowerCamelCase , __lowerCamelCase : int = evaluate(args) logger.info(f"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
38
from __future__ import annotations __lowerCamelCase : Optional[int] = """Muhammad Umer Farooq""" __lowerCamelCase : Tuple = """MIT""" __lowerCamelCase : Optional[int] = """1.0.0""" __lowerCamelCase : int = """Muhammad Umer Farooq""" __lowerCamelCase : Optional[int] = """contact@muhammadumerfarooq.me""" __lowerCamelCase : Dict = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : list[str] = [] UpperCamelCase : str = domain def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCamelCase : Any = parse.urljoin(self.domain , A_ ) self.urls.append(A_ ) def A_ ( _lowerCAmelCase ) -> str: return ".".join(get_sub_domain_name(_lowerCAmelCase ).split("." )[-2:] ) def A_ ( _lowerCAmelCase ) -> str: return parse.urlparse(_lowerCAmelCase ).netloc def A_ ( _lowerCAmelCase = "https://github.com" ) -> list[str]: UpperCamelCase : int = get_domain_name(_lowerCAmelCase ) # Initialize the parser UpperCamelCase : str = Parser(_lowerCAmelCase ) try: # Open URL UpperCamelCase : int = requests.get(_lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCamelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCamelCase : Optional[Any] = requests.get(_lowerCAmelCase ) # Get the valid email. UpperCamelCase : Optional[int] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = emails_from_url("""https://github.com""") print(f"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
38
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: __lowerCamelCase : Optional[int] = None __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : List[Any] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} __lowerCamelCase : Tuple = { """vocab_file""": { """facebook/mbart-large-en-ro""": ( """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model""" ), """facebook/mbart-large-cc25""": ( """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/mbart-large-en-ro""": """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json""", """facebook/mbart-large-cc25""": """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json""", }, } __lowerCamelCase : Optional[int] = { """facebook/mbart-large-en-ro""": 1024, """facebook/mbart-large-cc25""": 1024, } # fmt: off __lowerCamelCase : Any = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN"""] class A__ ( __snake_case ): _UpperCAmelCase :Tuple = VOCAB_FILES_NAMES _UpperCAmelCase :Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :Any = ['input_ids', 'attention_mask'] _UpperCAmelCase :str = MBartTokenizer _UpperCAmelCase :List[int] = [] _UpperCAmelCase :List[int] = [] def __init__( self , A_=None , A_=None , A_="<s>" , A_="</s>" , A_="</s>" , A_="<s>" , A_="<unk>" , A_="<pad>" , A_="<mask>" , A_=None , A_=None , A_=None , **A_ , ): '''simple docstring''' UpperCamelCase : int = AddedToken(A_ , lstrip=A_ , rstrip=A_ ) if isinstance(A_ , A_ ) else mask_token super().__init__( vocab_file=A_ , tokenizer_file=A_ , bos_token=A_ , eos_token=A_ , sep_token=A_ , cls_token=A_ , unk_token=A_ , pad_token=A_ , mask_token=A_ , src_lang=A_ , tgt_lang=A_ , additional_special_tokens=A_ , **A_ , ) UpperCamelCase : int = vocab_file UpperCamelCase : int = False if not self.vocab_file else True UpperCamelCase : Optional[int] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) UpperCamelCase : Optional[int] = { lang_code: self.convert_tokens_to_ids(A_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } UpperCamelCase : Optional[Any] = src_lang if src_lang is not None else "en_XX" UpperCamelCase : List[str] = self.convert_tokens_to_ids(self._src_lang ) UpperCamelCase : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __UpperCamelCase( self ): '''simple docstring''' return self._src_lang @src_lang.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __UpperCamelCase( self , A_ , A_ , A_ , A_ , **A_ ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCamelCase : Tuple = src_lang UpperCamelCase : Union[str, Any] = self(A_ , add_special_tokens=A_ , return_tensors=A_ , **A_ ) UpperCamelCase : int = self.convert_tokens_to_ids(A_ ) UpperCamelCase : Tuple = tgt_lang_id return inputs def __UpperCamelCase( self , A_ , A_ = "en_XX" , A_ = None , A_ = "ro_RO" , **A_ , ): '''simple docstring''' UpperCamelCase : Dict = src_lang UpperCamelCase : Optional[int] = tgt_lang return super().prepare_seqaseq_batch(A_ , A_ , **A_ ) def __UpperCamelCase( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def __UpperCamelCase( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = self.convert_tokens_to_ids(A_ ) UpperCamelCase : Union[str, Any] = [] UpperCamelCase : List[str] = [self.eos_token_id, self.cur_lang_code] UpperCamelCase : Dict = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCamelCase : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCamelCase : Dict = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.convert_tokens_to_ids(A_ ) UpperCamelCase : List[Any] = [] UpperCamelCase : Tuple = [self.eos_token_id, self.cur_lang_code] UpperCamelCase : Dict = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCamelCase : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCamelCase : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(A_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return UpperCamelCase : Optional[Any] = os.path.join( A_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ): copyfile(self.vocab_file , A_ ) return (out_vocab_file,)
38
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> list[int]: UpperCamelCase : Optional[Any] = [True] * limit UpperCamelCase : Optional[Any] = False UpperCamelCase : List[str] = False UpperCamelCase : Tuple = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCamelCase : Optional[Any] = i * 2 while index < limit: UpperCamelCase : int = False UpperCamelCase : Optional[int] = index + i UpperCamelCase : Any = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def A_ ( _lowerCAmelCase = 100_0000 ) -> int: UpperCamelCase : Union[str, Any] = prime_sieve(_lowerCAmelCase ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCamelCase : Dict = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCamelCase : int = j - i UpperCamelCase : Dict = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
38
1
import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("""1.0.0a"""): raise Exception("""requires fairseq >= 1.0.0a""") logging.set_verbosity_info() __lowerCamelCase : List[Any] = logging.get_logger(__name__) __lowerCamelCase : Tuple = """Hello world! cécé herlolip""" def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: UpperCamelCase : List[Any] = FairseqRobertaModel.from_pretrained(_lowerCAmelCase ) roberta.eval() # disable dropout UpperCamelCase : Any = roberta.model.encoder.sentence_encoder UpperCamelCase : int = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1e-5 , ) if classification_head: UpperCamelCase : Dict = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" , _lowerCAmelCase ) UpperCamelCase : Any = XLMRobertaXLForSequenceClassification(_lowerCAmelCase ) if classification_head else XLMRobertaXLForMaskedLM(_lowerCAmelCase ) model.eval() # Now let's copy all the weights. # Embeddings UpperCamelCase : List[Any] = roberta_sent_encoder.embed_tokens.weight UpperCamelCase : Dict = roberta_sent_encoder.embed_positions.weight UpperCamelCase : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. UpperCamelCase : Tuple = roberta_sent_encoder.layer_norm.weight UpperCamelCase : Union[str, Any] = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCamelCase : BertLayer = model.roberta.encoder.layer[i] UpperCamelCase : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] UpperCamelCase : RobertaAttention = layer.attention UpperCamelCase : int = roberta_layer.self_attn_layer_norm.weight UpperCamelCase : Union[str, Any] = roberta_layer.self_attn_layer_norm.bias # self attention UpperCamelCase : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) UpperCamelCase : Dict = roberta_layer.self_attn.q_proj.weight UpperCamelCase : Optional[int] = roberta_layer.self_attn.q_proj.bias UpperCamelCase : Optional[Any] = roberta_layer.self_attn.k_proj.weight UpperCamelCase : int = roberta_layer.self_attn.k_proj.bias UpperCamelCase : Optional[int] = roberta_layer.self_attn.v_proj.weight UpperCamelCase : Any = roberta_layer.self_attn.v_proj.bias # self-attention output UpperCamelCase : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape UpperCamelCase : str = roberta_layer.self_attn.out_proj.weight UpperCamelCase : Tuple = roberta_layer.self_attn.out_proj.bias # this one is final layer norm UpperCamelCase : Any = roberta_layer.final_layer_norm.weight UpperCamelCase : Dict = roberta_layer.final_layer_norm.bias # intermediate UpperCamelCase : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape UpperCamelCase : List[str] = roberta_layer.fca.weight UpperCamelCase : Union[str, Any] = roberta_layer.fca.bias # output UpperCamelCase : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape UpperCamelCase : str = roberta_layer.fca.weight UpperCamelCase : Dict = roberta_layer.fca.bias # end of layer if classification_head: UpperCamelCase : Dict = roberta.model.classification_heads["mnli"].dense.weight UpperCamelCase : List[str] = roberta.model.classification_heads["mnli"].dense.bias UpperCamelCase : int = roberta.model.classification_heads["mnli"].out_proj.weight UpperCamelCase : Dict = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head UpperCamelCase : Union[str, Any] = roberta.model.encoder.lm_head.dense.weight UpperCamelCase : Union[str, Any] = roberta.model.encoder.lm_head.dense.bias UpperCamelCase : Tuple = roberta.model.encoder.lm_head.layer_norm.weight UpperCamelCase : Tuple = roberta.model.encoder.lm_head.layer_norm.bias UpperCamelCase : Optional[Any] = roberta.model.encoder.lm_head.weight UpperCamelCase : Tuple = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCamelCase : torch.Tensor = roberta.encode(_lowerCAmelCase ).unsqueeze(0 ) # batch of size 1 UpperCamelCase : str = model(_lowerCAmelCase )[0] if classification_head: UpperCamelCase : str = roberta.model.classification_heads["mnli"](roberta.extract_features(_lowerCAmelCase ) ) else: UpperCamelCase : Union[str, Any] = roberta.model(_lowerCAmelCase )[0] print(our_output.shape , their_output.shape ) UpperCamelCase : Dict = torch.max(torch.abs(our_output - their_output ) ).item() print(F"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 UpperCamelCase : str = torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ) print("Do both models output the same tensors?" , "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(_lowerCAmelCase ).mkdir(parents=_lowerCAmelCase , exist_ok=_lowerCAmelCase ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--roberta_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--classification_head""", action="""store_true""", help="""Whether to convert a final classification head.""" ) __lowerCamelCase : Dict = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
38
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __snake_case ): def __init__( self , A_ , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , A_ = None , **A_ , ): '''simple docstring''' super().__init__( features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , num_proc=A_ , **A_ , ) UpperCamelCase : Optional[int] = Generator( cache_dir=A_ , features=A_ , generator=A_ , gen_kwargs=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' if self.streaming: UpperCamelCase : Optional[Any] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: UpperCamelCase : Union[str, Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , num_proc=self.num_proc , ) UpperCamelCase : int = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset
38
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class A__ ( __snake_case ): @staticmethod @abstractmethod def __UpperCamelCase( A_ ): '''simple docstring''' raise NotImplementedError() @abstractmethod def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError()
38
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
1
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCamelCase : str = {"""tokenization_tapex""": ["""TapexTokenizer"""]} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
38
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
1
import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput __lowerCamelCase : Dict = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class A__ ( __snake_case ): def __init__( self , *A_ , A_=None , A_=None , A_=None , **A_ ): '''simple docstring''' super().__init__(*A_ , **A_ ) UpperCamelCase : Tuple = eval_examples UpperCamelCase : Dict = post_process_function UpperCamelCase : Union[str, Any] = quant_trainer_args UpperCamelCase : Any = 128 # default number of calibration samples def __UpperCamelCase( self , A_=None ): '''simple docstring''' if calib_dataset is None and self.calib_dataset is None: raise ValueError("Trainer: calibration requires an calib_dataset." ) UpperCamelCase : Any = calib_dataset if calib_dataset is not None else self.calib_dataset UpperCamelCase : Dict = self._remove_unused_columns(A_ , description="Calibration" ) return DataLoader( A_ , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=A_ , ) def __UpperCamelCase( self , A_=None ): '''simple docstring''' UpperCamelCase : Any = self.train_dataset if calib_dataset is None else calib_dataset UpperCamelCase : Dict = self.get_calib_dataloader(A_ ) UpperCamelCase : Optional[int] = self.model quant_trainer.configure_model(A_ , self.quant_trainer_args , calib=A_ ) model.eval() quant_trainer.enable_calibration(A_ ) logger.info("***** Running calibration *****" ) logger.info(F""" Num examples = {self.calib_num}""" ) logger.info(F""" Batch size = {calib_dataloader.batch_size}""" ) for step, inputs in enumerate(A_ ): # Prediction step UpperCamelCase , UpperCamelCase , UpperCamelCase : List[Any] = self.prediction_step(A_ , A_ , prediction_loss_only=A_ ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(A_ , self.quant_trainer_args ) UpperCamelCase : str = model def __UpperCamelCase( self , A_=None , A_=None , A_=None , A_ = "eval" ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.eval_dataset if eval_dataset is None else eval_dataset UpperCamelCase : Any = self.get_eval_dataloader(A_ ) UpperCamelCase : List[Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. UpperCamelCase : List[Any] = self.compute_metrics UpperCamelCase : Dict = None UpperCamelCase : Any = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCamelCase : int = eval_loop( A_ , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=A_ , ) finally: UpperCamelCase : Union[str, Any] = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: UpperCamelCase : List[str] = self.post_process_function(A_ , A_ , output.predictions ) UpperCamelCase : Dict = self.compute_metrics(A_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): UpperCamelCase : Optional[int] = metrics.pop(A_ ) self.log(A_ ) else: UpperCamelCase : List[Any] = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) UpperCamelCase : Tuple = self.callback_handler.on_evaluate(self.args , self.state , self.control , A_ ) return metrics def __UpperCamelCase( self , A_ , A_ , A_=None , A_ = "test" ): '''simple docstring''' UpperCamelCase : Optional[int] = self.get_test_dataloader(A_ ) # Temporarily disable metric computation, we will do it in the loop here. UpperCamelCase : int = self.compute_metrics UpperCamelCase : Tuple = None UpperCamelCase : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCamelCase : Any = eval_loop( A_ , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=A_ , ) finally: UpperCamelCase : Dict = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output UpperCamelCase : List[str] = self.post_process_function(A_ , A_ , output.predictions , "predict" ) UpperCamelCase : Optional[Any] = self.compute_metrics(A_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): UpperCamelCase : Tuple = metrics.pop(A_ ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=A_ ) def __UpperCamelCase( self , A_="./" ): '''simple docstring''' UpperCamelCase : str = self.eval_dataset UpperCamelCase : Dict = self.get_eval_dataloader(A_ ) UpperCamelCase : List[str] = next(iter(A_ ) ) # saving device - to make it consistent UpperCamelCase : Tuple = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) # convert to tuple UpperCamelCase : Union[str, Any] = tuple(v.to(A_ ) for k, v in batch.items() ) logger.info("Converting model to be onnx compatible" ) from pytorch_quantization.nn import TensorQuantizer UpperCamelCase : str = True UpperCamelCase : int = self.model.to(A_ ) model.eval() model.float() UpperCamelCase : Tuple = model.module if hasattr(A_ , "module" ) else model quant_trainer.configure_model(A_ , self.quant_trainer_args ) UpperCamelCase : int = os.path.join(A_ , "model.onnx" ) logger.info(F"""exporting model to {output_model_file}""" ) UpperCamelCase : List[Any] = {0: "batch_size", 1: "seq_len"} torch.onnx.export( A_ , A_ , A_ , export_params=A_ , opset_version=13 , do_constant_folding=A_ , input_names=["input_ids", "attention_mask", "token_type_ids"] , output_names=["output_start_logits", "output_end_logits"] , dynamic_axes={ "input_ids": axes, "attention_mask": axes, "token_type_ids": axes, "output_start_logits": axes, "output_end_logits": axes, } , verbose=A_ , ) logger.info("onnx export finished" )
38
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :int = CodeGenTokenizer _UpperCAmelCase :Dict = CodeGenTokenizerFast _UpperCAmelCase :Any = True _UpperCAmelCase :Any = {'add_prefix_space': True} _UpperCAmelCase :Optional[int] = False def __UpperCamelCase( self ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase : Dict = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] UpperCamelCase : List[str] = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase : Union[str, Any] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCamelCase : Optional[int] = {"unk_token": "<unk>"} UpperCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(A_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(A_ ) ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Tuple = "lower newer" UpperCamelCase : Dict = "lower newer" return input_text, output_text def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase : Tuple = "lower newer" UpperCamelCase : Optional[Any] = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] UpperCamelCase : List[Any] = tokenizer.tokenize(A_ , add_prefix_space=A_ ) self.assertListEqual(A_ , A_ ) UpperCamelCase : Dict = tokens + [tokenizer.unk_token] UpperCamelCase : Optional[Any] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' if not self.test_rust_tokenizer: return UpperCamelCase : Any = self.get_tokenizer() UpperCamelCase : int = self.get_rust_tokenizer(add_prefix_space=A_ ) UpperCamelCase : Dict = "lower newer" # Testing tokenization UpperCamelCase : Tuple = tokenizer.tokenize(A_ , add_prefix_space=A_ ) UpperCamelCase : int = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids without special tokens UpperCamelCase : Optional[int] = tokenizer.encode(A_ , add_special_tokens=A_ , add_prefix_space=A_ ) UpperCamelCase : Tuple = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids with special tokens UpperCamelCase : List[str] = self.get_rust_tokenizer(add_prefix_space=A_ ) UpperCamelCase : Tuple = tokenizer.encode(A_ , add_prefix_space=A_ ) UpperCamelCase : List[Any] = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # Testing the unknown token UpperCamelCase : Any = tokens + [rust_tokenizer.unk_token] UpperCamelCase : str = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' pass def __UpperCamelCase( self , A_=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCamelCase : Any = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) # Simple input UpperCamelCase : Any = "This is a simple input" UpperCamelCase : Optional[int] = ["This is a simple input 1", "This is a simple input 2"] UpperCamelCase : Optional[int] = ("This is a simple input", "This is a pair") UpperCamelCase : Dict = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding="max_length" ) # Simple input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding="max_length" ) # Simple input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding="max_length" , ) # Pair input self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding="max_length" ) # Pair input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding="max_length" ) # Pair input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding="max_length" , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>" ) # Simple input UpperCamelCase : Optional[int] = "This is a simple input" UpperCamelCase : Dict = ["This is a simple input looooooooong", "This is a simple input"] UpperCamelCase : Optional[int] = ("This is a simple input", "This is a pair") UpperCamelCase : Optional[Any] = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] UpperCamelCase : int = tokenizer.pad_token_id UpperCamelCase : List[str] = tokenizer(A_ , padding="max_length" , max_length=30 , return_tensors="np" ) UpperCamelCase : List[Any] = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors="np" ) UpperCamelCase : List[str] = tokenizer(*A_ , padding="max_length" , max_length=60 , return_tensors="np" ) UpperCamelCase : Union[str, Any] = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors="np" ) # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s["input_ids"] ) self.assertTrue(0 in out_s["attention_mask"] ) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0] ) self.assertFalse(0 in out_sa["attention_mask"][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1] ) self.assertTrue(0 in out_sa["attention_mask"][1] ) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p["input_ids"] ) self.assertTrue(0 in out_p["attention_mask"] ) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0] ) self.assertFalse(0 in out_pa["attention_mask"][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1] ) self.assertTrue(0 in out_pa["attention_mask"][1] ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = "$$$" UpperCamelCase : str = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A_ , add_bos_token=A_ ) UpperCamelCase : Tuple = "This is a simple input" UpperCamelCase : str = ["This is a simple input 1", "This is a simple input 2"] UpperCamelCase : Tuple = tokenizer.bos_token_id UpperCamelCase : Union[str, Any] = tokenizer(A_ ) UpperCamelCase : int = tokenizer(A_ ) self.assertEqual(out_s.input_ids[0] , A_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) UpperCamelCase : int = tokenizer.decode(out_s.input_ids ) UpperCamelCase : Optional[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = CodeGenTokenizer.from_pretrained("Salesforce/codegen-350M-mono" ) UpperCamelCase : List[str] = "\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#" UpperCamelCase : Optional[int] = "\nif len_a > len_b: result = a\nelse: result = b" UpperCamelCase : int = tokenizer.encode(A_ ) UpperCamelCase : int = ["^#", re.escape("<|endoftext|>" ), "^'''", "^\"\"\"", "\n\n\n"] UpperCamelCase : Any = tokenizer.decode(A_ , truncate_before_pattern=A_ ) self.assertEqual(A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' pass
38
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
1
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowerCamelCase : Any = logging.get_logger(__name__) __lowerCamelCase : int = { """facebook/detr-resnet-50""": """https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json""", # See all DETR models at https://huggingface.co/models?filter=detr } class A__ ( __snake_case ): _UpperCAmelCase :int = 'detr' _UpperCAmelCase :List[Any] = ['past_key_values'] _UpperCAmelCase :Optional[int] = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCamelCase : Optional[int] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(A_ , A_ ): UpperCamelCase : Tuple = backbone_config.get("model_type" ) UpperCamelCase : Optional[int] = CONFIG_MAPPING[backbone_model_type] UpperCamelCase : List[str] = config_class.from_dict(A_ ) # set timm attributes to None UpperCamelCase , UpperCamelCase , UpperCamelCase : Optional[Any] = None, None, None UpperCamelCase : Optional[int] = use_timm_backbone UpperCamelCase : Any = backbone_config UpperCamelCase : Union[str, Any] = num_channels UpperCamelCase : int = num_queries UpperCamelCase : Union[str, Any] = d_model UpperCamelCase : List[Any] = encoder_ffn_dim UpperCamelCase : Optional[int] = encoder_layers UpperCamelCase : List[str] = encoder_attention_heads UpperCamelCase : int = decoder_ffn_dim UpperCamelCase : Optional[Any] = decoder_layers UpperCamelCase : List[str] = decoder_attention_heads UpperCamelCase : int = dropout UpperCamelCase : List[str] = attention_dropout UpperCamelCase : int = activation_dropout UpperCamelCase : Dict = activation_function UpperCamelCase : Any = init_std UpperCamelCase : Any = init_xavier_std UpperCamelCase : Dict = encoder_layerdrop UpperCamelCase : int = decoder_layerdrop UpperCamelCase : Tuple = encoder_layers UpperCamelCase : List[str] = auxiliary_loss UpperCamelCase : List[str] = position_embedding_type UpperCamelCase : str = backbone UpperCamelCase : str = use_pretrained_backbone UpperCamelCase : Optional[Any] = dilation # Hungarian matcher UpperCamelCase : Union[str, Any] = class_cost UpperCamelCase : Optional[int] = bbox_cost UpperCamelCase : Union[str, Any] = giou_cost # Loss coefficients UpperCamelCase : List[str] = mask_loss_coefficient UpperCamelCase : Optional[int] = dice_loss_coefficient UpperCamelCase : Optional[int] = bbox_loss_coefficient UpperCamelCase : str = giou_loss_coefficient UpperCamelCase : str = eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.encoder_attention_heads @property def __UpperCamelCase( self ): '''simple docstring''' return self.d_model @classmethod def __UpperCamelCase( cls , A_ , **A_ ): '''simple docstring''' return cls(backbone_config=A_ , **A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCamelCase : Tuple = self.backbone_config.to_dict() UpperCamelCase : int = self.__class__.model_type return output class A__ ( __snake_case ): _UpperCAmelCase :List[Any] = version.parse('1.11' ) @property def __UpperCamelCase( self ): '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def __UpperCamelCase( self ): '''simple docstring''' return 1e-5 @property def __UpperCamelCase( self ): '''simple docstring''' return 12
38
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A_ ( ) -> Dict: UpperCamelCase : Tuple = ArgumentParser( description=( "PyTorch TPU distributed training launch " "helper utility that will spawn up " "multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCAmelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCAmelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCAmelCase ) return parser.parse_args() def A_ ( ) -> Optional[int]: UpperCamelCase : Tuple = parse_args() # Import training_script as a module. UpperCamelCase : Union[str, Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase : List[Any] = script_fpath.stem UpperCamelCase : Optional[Any] = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv UpperCamelCase : List[Any] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
38
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class A__ ( __snake_case ): _UpperCAmelCase :str = ['image_processor'] _UpperCAmelCase :Tuple = 'SamImageProcessor' def __init__( self , A_ ): '''simple docstring''' super().__init__(A_ ) UpperCamelCase : List[Any] = self.image_processor UpperCamelCase : Tuple = -10 UpperCamelCase : Optional[int] = self.image_processor.size["longest_edge"] def __call__( self , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ): '''simple docstring''' UpperCamelCase : Dict = self.image_processor( A_ , return_tensors=A_ , **A_ , ) # pop arguments that are not used in the foward but used nevertheless UpperCamelCase : Union[str, Any] = encoding_image_processor["original_sizes"] if hasattr(A_ , "numpy" ): # Checks if Torch or TF tensor UpperCamelCase : Tuple = original_sizes.numpy() UpperCamelCase , UpperCamelCase , UpperCamelCase : Dict = self._check_and_preprocess_points( input_points=A_ , input_labels=A_ , input_boxes=A_ , ) UpperCamelCase : List[str] = self._normalize_and_convert( A_ , A_ , input_points=A_ , input_labels=A_ , input_boxes=A_ , return_tensors=A_ , ) return encoding_image_processor def __UpperCamelCase( self , A_ , A_ , A_=None , A_=None , A_=None , A_="pt" , ): '''simple docstring''' if input_points is not None: if len(A_ ) != len(A_ ): UpperCamelCase : Union[str, Any] = [ self._normalize_coordinates(self.target_size , A_ , original_sizes[0] ) for point in input_points ] else: UpperCamelCase : Optional[int] = [ self._normalize_coordinates(self.target_size , A_ , A_ ) for point, original_size in zip(A_ , A_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: UpperCamelCase , UpperCamelCase : int = self._pad_points_and_labels(A_ , A_ ) UpperCamelCase : Optional[Any] = np.array(A_ ) if input_labels is not None: UpperCamelCase : Optional[Any] = np.array(A_ ) if input_boxes is not None: if len(A_ ) != len(A_ ): UpperCamelCase : Any = [ self._normalize_coordinates(self.target_size , A_ , original_sizes[0] , is_bounding_box=A_ ) for box in input_boxes ] else: UpperCamelCase : Dict = [ self._normalize_coordinates(self.target_size , A_ , A_ , is_bounding_box=A_ ) for box, original_size in zip(A_ , A_ ) ] UpperCamelCase : Dict = np.array(A_ ) if input_boxes is not None: if return_tensors == "pt": UpperCamelCase : Optional[int] = torch.from_numpy(A_ ) # boxes batch size of 1 by default UpperCamelCase : Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": UpperCamelCase : str = tf.convert_to_tensor(A_ ) # boxes batch size of 1 by default UpperCamelCase : Any = tf.expand_dims(A_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": UpperCamelCase : int = torch.from_numpy(A_ ) # point batch size of 1 by default UpperCamelCase : int = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": UpperCamelCase : Optional[int] = tf.convert_to_tensor(A_ ) # point batch size of 1 by default UpperCamelCase : Optional[Any] = tf.expand_dims(A_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": UpperCamelCase : str = torch.from_numpy(A_ ) # point batch size of 1 by default UpperCamelCase : Optional[int] = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": UpperCamelCase : Any = tf.convert_to_tensor(A_ ) # point batch size of 1 by default UpperCamelCase : Dict = tf.expand_dims(A_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = max([point.shape[0] for point in input_points] ) UpperCamelCase : List[Any] = [] for i, point in enumerate(A_ ): if point.shape[0] != expected_nb_points: UpperCamelCase : str = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) UpperCamelCase : Dict = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(A_ ) UpperCamelCase : Union[str, Any] = processed_input_points return input_points, input_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_=False ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = original_size UpperCamelCase , UpperCamelCase : str = self.image_processor._get_preprocess_shape(A_ , longest_edge=A_ ) UpperCamelCase : List[Any] = deepcopy(A_ ).astype(A_ ) if is_bounding_box: UpperCamelCase : Optional[Any] = coords.reshape(-1 , 2 , 2 ) UpperCamelCase : Dict = coords[..., 0] * (new_w / old_w) UpperCamelCase : int = coords[..., 1] * (new_h / old_h) if is_bounding_box: UpperCamelCase : str = coords.reshape(-1 , 4 ) return coords def __UpperCamelCase( self , A_=None , A_=None , A_=None , ): '''simple docstring''' if input_points is not None: if hasattr(A_ , "numpy" ): # Checks for TF or Torch tensor UpperCamelCase : List[Any] = input_points.numpy().tolist() if not isinstance(A_ , A_ ) or not isinstance(input_points[0] , A_ ): raise ValueError("Input points must be a list of list of floating points." ) UpperCamelCase : Any = [np.array(A_ ) for input_point in input_points] else: UpperCamelCase : Union[str, Any] = None if input_labels is not None: if hasattr(A_ , "numpy" ): UpperCamelCase : str = input_labels.numpy().tolist() if not isinstance(A_ , A_ ) or not isinstance(input_labels[0] , A_ ): raise ValueError("Input labels must be a list of list integers." ) UpperCamelCase : str = [np.array(A_ ) for label in input_labels] else: UpperCamelCase : Optional[Any] = None if input_boxes is not None: if hasattr(A_ , "numpy" ): UpperCamelCase : Union[str, Any] = input_boxes.numpy().tolist() if ( not isinstance(A_ , A_ ) or not isinstance(input_boxes[0] , A_ ) or not isinstance(input_boxes[0][0] , A_ ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) UpperCamelCase : Optional[int] = [np.array(A_ ).astype(np.floataa ) for box in input_boxes] else: UpperCamelCase : Dict = None return input_points, input_labels, input_boxes @property def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.image_processor.model_input_names return list(dict.fromkeys(A_ ) ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.image_processor.post_process_masks(*A_ , **A_ )
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
def A_ ( _lowerCAmelCase ) -> int: if not head: return True # split the list to two parts UpperCamelCase , UpperCamelCase : Dict = head.next, head while fast and fast.next: UpperCamelCase : Union[str, Any] = fast.next.next UpperCamelCase : Dict = slow.next UpperCamelCase : Optional[Any] = slow.next UpperCamelCase : str = None # Don't forget here! But forget still works! # reverse the second part UpperCamelCase : Optional[Any] = None while second: UpperCamelCase : Dict = second.next UpperCamelCase : List[str] = node UpperCamelCase : List[Any] = second UpperCamelCase : int = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False UpperCamelCase : List[str] = node.next UpperCamelCase : str = head.next return True def A_ ( _lowerCAmelCase ) -> Dict: if not head or not head.next: return True # 1. Get the midpoint (slow) UpperCamelCase : Optional[int] = head while fast and fast.next: UpperCamelCase , UpperCamelCase : Optional[Any] = fast.next.next, slow.next # 2. Push the second half into the stack UpperCamelCase : Optional[int] = [slow.val] while slow.next: UpperCamelCase : int = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False UpperCamelCase : Union[str, Any] = cur.next return True def A_ ( _lowerCAmelCase ) -> Any: if not head or not head.next: return True UpperCamelCase : List[Any] = {} UpperCamelCase : Any = 0 while head: if head.val in d: d[head.val].append(_lowerCAmelCase ) else: UpperCamelCase : Optional[int] = [pos] UpperCamelCase : Optional[int] = head.next pos += 1 UpperCamelCase : Any = pos - 1 UpperCamelCase : Dict = 0 for v in d.values(): if len(_lowerCAmelCase ) % 2 != 0: middle += 1 else: UpperCamelCase : str = 0 for i in range(0 , len(_lowerCAmelCase ) ): if v[i] + v[len(_lowerCAmelCase ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
38
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
1
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> float: if not nums: raise ValueError("List is empty" ) return sum(_lowerCAmelCase ) / len(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
38
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __lowerCamelCase : Dict = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __lowerCamelCase : Tuple = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = VOCAB_FILES_NAMES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[Any] = BlenderbotSmallTokenizer def __init__( self , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , A_=True , **A_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=A_ , merges=A_ , add_prefix_space=A_ , trim_offsets=A_ , ) , bos_token=A_ , eos_token=A_ , unk_token=A_ , **A_ , ) UpperCamelCase : Union[str, Any] = add_prefix_space def __UpperCamelCase( self , A_ , A_=None ): '''simple docstring''' UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
38
1
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class A__ ( __snake_case ): _UpperCAmelCase :str = (DPMSolverSinglestepScheduler,) _UpperCAmelCase :Any = (('num_inference_steps', 2_5),) def __UpperCamelCase( self , **A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", "solver_order": 2, "prediction_type": "epsilon", "thresholding": False, "sample_max_value": 1.0, "algorithm_type": "dpmsolver++", "solver_type": "midpoint", "lambda_min_clipped": -float("inf" ), "variance_type": None, } config.update(**A_ ) return config def __UpperCamelCase( self , A_=0 , **A_ ): '''simple docstring''' UpperCamelCase : str = dict(self.forward_default_kwargs ) UpperCamelCase : Dict = kwargs.pop("num_inference_steps" , A_ ) UpperCamelCase : Any = self.dummy_sample UpperCamelCase : Union[str, Any] = 0.1 * sample UpperCamelCase : int = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: UpperCamelCase : int = self.get_scheduler_config(**A_ ) UpperCamelCase : str = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals UpperCamelCase : Dict = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) UpperCamelCase : List[Any] = scheduler_class.from_pretrained(A_ ) new_scheduler.set_timesteps(A_ ) # copy over dummy past residuals UpperCamelCase : Any = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCamelCase , UpperCamelCase : str = sample, sample for t in range(A_ , time_step + scheduler.config.solver_order + 1 ): UpperCamelCase : Union[str, Any] = scheduler.step(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : List[Any] = new_scheduler.step(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self , A_=0 , **A_ ): '''simple docstring''' UpperCamelCase : Tuple = dict(self.forward_default_kwargs ) UpperCamelCase : Dict = kwargs.pop("num_inference_steps" , A_ ) UpperCamelCase : int = self.dummy_sample UpperCamelCase : Dict = 0.1 * sample UpperCamelCase : Any = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: UpperCamelCase : Union[str, Any] = self.get_scheduler_config() UpperCamelCase : Dict = scheduler_class(**A_ ) scheduler.set_timesteps(A_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCamelCase : Tuple = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A_ ) UpperCamelCase : Dict = scheduler_class.from_pretrained(A_ ) # copy over dummy past residuals new_scheduler.set_timesteps(A_ ) # copy over dummy past residual (must be after setting timesteps) UpperCamelCase : Dict = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCamelCase : Tuple = scheduler.step(A_ , A_ , A_ , **A_ ).prev_sample UpperCamelCase : str = new_scheduler.step(A_ , A_ , A_ , **A_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase( self , A_=None , **A_ ): '''simple docstring''' if scheduler is None: UpperCamelCase : Any = self.scheduler_classes[0] UpperCamelCase : Any = self.get_scheduler_config(**A_ ) UpperCamelCase : Union[str, Any] = scheduler_class(**A_ ) UpperCamelCase : Any = self.scheduler_classes[0] UpperCamelCase : Dict = self.get_scheduler_config(**A_ ) UpperCamelCase : Union[str, Any] = scheduler_class(**A_ ) UpperCamelCase : Optional[Any] = 10 UpperCamelCase : Union[str, Any] = self.dummy_model() UpperCamelCase : int = self.dummy_sample_deter scheduler.set_timesteps(A_ ) for i, t in enumerate(scheduler.timesteps ): UpperCamelCase : Optional[int] = model(A_ , A_ ) UpperCamelCase : Any = scheduler.step(A_ , A_ , A_ ).prev_sample return sample def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) UpperCamelCase : List[str] = 50 UpperCamelCase : Tuple = self.dummy_model() UpperCamelCase : Dict = self.dummy_sample_deter scheduler.set_timesteps(A_ ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): UpperCamelCase : List[Any] = model(A_ , A_ ) UpperCamelCase : List[Any] = scheduler.step(A_ , A_ , A_ ).prev_sample UpperCamelCase : List[str] = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.25_74 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) UpperCamelCase : Optional[int] = self.full_loop(scheduler=A_ ) UpperCamelCase : Tuple = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1e-3 UpperCamelCase : Union[str, Any] = DEISMultistepScheduler.from_config(scheduler.config ) UpperCamelCase : Tuple = DPMSolverMultistepScheduler.from_config(scheduler.config ) UpperCamelCase : List[str] = UniPCMultistepScheduler.from_config(scheduler.config ) UpperCamelCase : Dict = DPMSolverSinglestepScheduler.from_config(scheduler.config ) UpperCamelCase : Optional[Any] = self.full_loop(scheduler=A_ ) UpperCamelCase : Optional[Any] = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' self.check_over_configs(thresholding=A_ ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A_ , prediction_type=A_ , sample_max_value=A_ , algorithm_type="dpmsolver++" , solver_order=A_ , solver_type=A_ , ) def __UpperCamelCase( self ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_ ) def __UpperCamelCase( self ): '''simple docstring''' for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A_ , solver_type=A_ , prediction_type=A_ , algorithm_type=A_ , ) UpperCamelCase : List[Any] = self.full_loop( solver_order=A_ , solver_type=A_ , prediction_type=A_ , algorithm_type=A_ , ) assert not torch.isnan(A_ ).any(), "Samples have nan numbers" def __UpperCamelCase( self ): '''simple docstring''' self.check_over_configs(lower_order_final=A_ ) self.check_over_configs(lower_order_final=A_ ) def __UpperCamelCase( self ): '''simple docstring''' self.check_over_configs(lambda_min_clipped=-float("inf" ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def __UpperCamelCase( self ): '''simple docstring''' self.check_over_configs(variance_type=A_ ) self.check_over_configs(variance_type="learned_range" ) def __UpperCamelCase( self ): '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=A_ , time_step=0 ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.full_loop() UpperCamelCase : Any = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.27_91 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.full_loop(use_karras_sigmas=A_ ) UpperCamelCase : List[Any] = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.22_48 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.full_loop(prediction_type="v_prediction" ) UpperCamelCase : Dict = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.14_53 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.full_loop(prediction_type="v_prediction" , use_karras_sigmas=A_ ) UpperCamelCase : Any = torch.mean(torch.abs(A_ ) ) assert abs(result_mean.item() - 0.06_49 ) < 1e-3 def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.scheduler_classes[0] UpperCamelCase : Optional[Any] = self.get_scheduler_config(thresholding=A_ , dynamic_thresholding_ratio=0 ) UpperCamelCase : Any = scheduler_class(**A_ ) UpperCamelCase : Optional[Any] = 10 UpperCamelCase : int = self.dummy_model() UpperCamelCase : List[Any] = self.dummy_sample_deter.half() scheduler.set_timesteps(A_ ) for i, t in enumerate(scheduler.timesteps ): UpperCamelCase : Optional[int] = model(A_ , A_ ) UpperCamelCase : List[str] = scheduler.step(A_ , A_ , A_ ).prev_sample assert sample.dtype == torch.floataa
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : int = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
import re from filelock import FileLock try: import nltk __lowerCamelCase : int = True except (ImportError, ModuleNotFoundError): __lowerCamelCase : Any = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def A_ ( _lowerCAmelCase ) -> str: re.sub("<n>" , "" , _lowerCAmelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(_lowerCAmelCase ) )
38
import logging import os import threading import time try: import warnings except ImportError: __lowerCamelCase : str = None try: import msvcrt except ImportError: __lowerCamelCase : str = None try: import fcntl except ImportError: __lowerCamelCase : List[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __lowerCamelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ __lowerCamelCase : str = [ """Timeout""", """BaseFileLock""", """WindowsFileLock""", """UnixFileLock""", """SoftFileLock""", """FileLock""", ] __lowerCamelCase : Union[str, Any] = """3.0.12""" __lowerCamelCase : Any = None def A_ ( ) -> List[Any]: global _logger UpperCamelCase : Any = _logger or logging.getLogger(__name__ ) return _logger class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = lock_file return None def __str__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = lock return None def __enter__( self ): '''simple docstring''' return self.lock def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.lock.release() return None class A__ : def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : List[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long UpperCamelCase : Dict = self.hash_filename_if_too_long(A_ , A_ ) # The path to the lock file. UpperCamelCase : List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. UpperCamelCase : Tuple = None # The default timeout value. UpperCamelCase : Optional[Any] = timeout # We use this lock primarily for the lock counter. UpperCamelCase : Union[str, Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. UpperCamelCase : Dict = 0 return None @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file @property def __UpperCamelCase( self ): '''simple docstring''' return self._timeout @timeout.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = float(A_ ) return None def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file_fd is not None def __UpperCamelCase( self , A_=None , A_=0.05 ): '''simple docstring''' if timeout is None: UpperCamelCase : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 UpperCamelCase : Dict = id(self ) UpperCamelCase : List[str] = self._lock_file UpperCamelCase : int = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(A_ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: UpperCamelCase : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __UpperCamelCase( self , A_=False ): '''simple docstring''' with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: UpperCamelCase : List[Any] = id(self ) UpperCamelCase : Dict = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() UpperCamelCase : Dict = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): '''simple docstring''' self.acquire() return self def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.release() return None def __del__( self ): '''simple docstring''' self.release(force=A_ ) return None def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = os.path.basename(A_ ) if len(A_ ) > max_length and max_length > 0: UpperCamelCase : Optional[int] = os.path.dirname(A_ ) UpperCamelCase : int = str(hash(A_ ) ) UpperCamelCase : Any = filename[: max_length - len(A_ ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(A_ , A_ ) else: return path class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' from .file_utils import relative_to_absolute_path super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) UpperCamelCase : List[Any] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: UpperCamelCase : str = os.open(self._lock_file , A_ ) except OSError: pass else: try: msvcrt.locking(A_ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(A_ ) else: UpperCamelCase : Optional[Any] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self._lock_file_fd UpperCamelCase : str = None msvcrt.locking(A_ , msvcrt.LK_UNLCK , 1 ) os.close(A_ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : Tuple = os.statvfs(os.path.dirname(A_ ) ).f_namemax super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC UpperCamelCase : int = os.open(self._lock_file , A_ ) try: fcntl.flock(A_ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(A_ ) else: UpperCamelCase : List[str] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self._lock_file_fd UpperCamelCase : List[Any] = None fcntl.flock(A_ , fcntl.LOCK_UN ) os.close(A_ ) return None class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: UpperCamelCase : Optional[int] = os.open(self._lock_file , A_ ) except OSError: pass else: UpperCamelCase : Tuple = fd return None def __UpperCamelCase( self ): '''simple docstring''' os.close(self._lock_file_fd ) UpperCamelCase : str = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __lowerCamelCase : Dict = None if msvcrt: __lowerCamelCase : Any = WindowsFileLock elif fcntl: __lowerCamelCase : Any = UnixFileLock else: __lowerCamelCase : int = SoftFileLock if warnings is not None: warnings.warn("""only soft file lock is available""")
38
1
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging __lowerCamelCase : Tuple = { """cola""": 2, """mnli""": 3, """mrpc""": 2, """sst-2""": 2, """sts-b""": 1, """qqp""": 2, """qnli""": 2, """rte""": 2, """wnli""": 2, } logging.set_verbosity_info() def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ) -> str: # Initialise PyTorch model UpperCamelCase : int = XLNetConfig.from_json_file(_lowerCAmelCase ) UpperCamelCase : List[str] = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) UpperCamelCase : Dict = finetuning_task UpperCamelCase : Any = GLUE_TASKS_NUM_LABELS[finetuning_task] UpperCamelCase : List[str] = XLNetForSequenceClassification(_lowerCAmelCase ) elif "squad" in finetuning_task: UpperCamelCase : List[Any] = finetuning_task UpperCamelCase : Optional[Any] = XLNetForQuestionAnswering(_lowerCAmelCase ) else: UpperCamelCase : Optional[int] = XLNetLMHeadModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model UpperCamelCase : List[str] = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : str = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) print(F"""Save PyTorch model to {os.path.abspath(_lowerCAmelCase )}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) print(F"""Save configuration file to {os.path.abspath(_lowerCAmelCase )}""" ) with open(_lowerCAmelCase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--xlnet_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained XLNet model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--finetuning_task""", default=None, type=str, help="""Name of a task on which the XLNet TensorFlow model was fine-tuned""", ) __lowerCamelCase : Dict = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
38
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> str: if config_name_or_path is None: UpperCamelCase : Dict = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCamelCase : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCamelCase : Tuple = question_encoder_name_or_path UpperCamelCase : Any = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCamelCase : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Tuple = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : int = gen_config UpperCamelCase : Dict = question_encoder_config UpperCamelCase : Tuple = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--model_type""", choices=["""rag_sequence""", """rag_token"""], required=True, type=str, help="""RAG model type: rag_sequence, rag_token""", ) parser.add_argument("""--dest""", type=str, required=True, help="""Path to the output checkpoint directory.""") parser.add_argument("""--generator_name_or_path""", type=str, required=True, help="""Generator model identifier""") parser.add_argument( """--question_encoder_name_or_path""", type=str, required=True, help="""Question encoder model identifier""" ) parser.add_argument( """--generator_tokenizer_name_or_path""", type=str, help="""Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``""", ) parser.add_argument( """--question_encoder_tokenizer_name_or_path""", type=str, help="""Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``""", ) parser.add_argument( """--config_name_or_path""", type=str, help=( """Identifier of the model config to use, if not provided, resolves to a base config for a given""" """ ``model_type``""" ), ) __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
38
1
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) __lowerCamelCase : Optional[Any] = logging.getLogger(__name__) @dataclass(frozen=__snake_case ) class A__ : _UpperCAmelCase :str _UpperCAmelCase :str _UpperCAmelCase :Optional[str] = None _UpperCAmelCase :Optional[str] = None _UpperCAmelCase :Optional[str] = None @dataclass(frozen=__snake_case ) class A__ : _UpperCAmelCase :List[int] _UpperCAmelCase :Optional[List[int]] = None _UpperCAmelCase :Optional[List[int]] = None _UpperCAmelCase :Optional[Union[int, float]] = None _UpperCAmelCase :Optional[int] = None if is_torch_available(): import torch from torch.utils.data import Dataset class A__ ( __snake_case ): _UpperCAmelCase :List[InputFeatures] def __init__( self , A_ , A_ , A_ , A_ = None , A_=False , A_ = False , ): '''simple docstring''' UpperCamelCase : Optional[Any] = hans_processors[task]() UpperCamelCase : Optional[Any] = os.path.join( A_ , "cached_{}_{}_{}_{}".format( "dev" if evaluate else "train" , tokenizer.__class__.__name__ , str(A_ ) , A_ , ) , ) UpperCamelCase : List[str] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCamelCase , UpperCamelCase : int = label_list[2], label_list[1] UpperCamelCase : Dict = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. UpperCamelCase : int = cached_features_file + ".lock" with FileLock(A_ ): if os.path.exists(A_ ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) UpperCamelCase : List[str] = torch.load(A_ ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) UpperCamelCase : List[Any] = ( processor.get_dev_examples(A_ ) if evaluate else processor.get_train_examples(A_ ) ) logger.info("Training examples: %s" , len(A_ ) ) UpperCamelCase : Tuple = hans_convert_examples_to_features(A_ , A_ , A_ , A_ ) logger.info("Saving features into cached file %s" , A_ ) torch.save(self.features , A_ ) def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , A_ ): '''simple docstring''' return self.features[i] def __UpperCamelCase( self ): '''simple docstring''' return self.label_list if is_tf_available(): import tensorflow as tf class A__ : _UpperCAmelCase :List[InputFeatures] def __init__( self , A_ , A_ , A_ , A_ = 128 , A_=False , A_ = False , ): '''simple docstring''' UpperCamelCase : Optional[Any] = hans_processors[task]() UpperCamelCase : Optional[int] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCamelCase , UpperCamelCase : Any = label_list[2], label_list[1] UpperCamelCase : Tuple = label_list UpperCamelCase : Optional[Any] = processor.get_dev_examples(A_ ) if evaluate else processor.get_train_examples(A_ ) UpperCamelCase : str = hans_convert_examples_to_features(A_ , A_ , A_ , A_ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc="convert examples to features" ): if ex_index % 1_0000 == 0: logger.info("Writing example %d of %d" % (ex_index, len(A_ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) UpperCamelCase : str = tf.data.Dataset.from_generator( A_ , ( { "example_id": tf.intaa, "input_ids": tf.intaa, "attention_mask": tf.intaa, "token_type_ids": tf.intaa, }, tf.intaa, ) , ( { "example_id": tf.TensorShape([] ), "input_ids": tf.TensorShape([None, None] ), "attention_mask": tf.TensorShape([None, None] ), "token_type_ids": tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def __UpperCamelCase( self ): '''simple docstring''' return self.dataset def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , A_ ): '''simple docstring''' return self.features[i] def __UpperCamelCase( self ): '''simple docstring''' return self.label_list class A__ ( __snake_case ): def __UpperCamelCase( self , A_ ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(A_ , "heuristics_train_set.txt" ) ) , "train" ) def __UpperCamelCase( self , A_ ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(A_ , "heuristics_evaluation_set.txt" ) ) , "dev" ) def __UpperCamelCase( self ): '''simple docstring''' return ["contradiction", "entailment", "neutral"] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = [] for i, line in enumerate(A_ ): if i == 0: continue UpperCamelCase : Union[str, Any] = "%s-%s" % (set_type, line[0]) UpperCamelCase : Optional[int] = line[5] UpperCamelCase : Tuple = line[6] UpperCamelCase : int = line[7][2:] if line[7].startswith("ex" ) else line[7] UpperCamelCase : Optional[Any] = line[0] examples.append(InputExample(guid=A_ , text_a=A_ , text_b=A_ , label=A_ , pairID=A_ ) ) return examples def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> int: UpperCamelCase : Optional[int] = {label: i for i, label in enumerate(_lowerCAmelCase )} UpperCamelCase : List[str] = [] for ex_index, example in tqdm.tqdm(enumerate(_lowerCAmelCase ) , desc="convert examples to features" ): if ex_index % 1_0000 == 0: logger.info("Writing example %d" % (ex_index) ) UpperCamelCase : List[str] = tokenizer( example.text_a , example.text_b , add_special_tokens=_lowerCAmelCase , max_length=_lowerCAmelCase , padding="max_length" , truncation=_lowerCAmelCase , return_overflowing_tokens=_lowerCAmelCase , ) UpperCamelCase : str = label_map[example.label] if example.label in label_map else 0 UpperCamelCase : Tuple = int(example.pairID ) features.append(InputFeatures(**_lowerCAmelCase , label=_lowerCAmelCase , pairID=_lowerCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info("*** Example ***" ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features __lowerCamelCase : Tuple = { """hans""": 3, } __lowerCamelCase : List[Any] = { """hans""": HansProcessor, }
38
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : str = 13 UpperCamelCase : int = 7 UpperCamelCase : str = True UpperCamelCase : Dict = True UpperCamelCase : str = True UpperCamelCase : Tuple = True UpperCamelCase : List[str] = 99 UpperCamelCase : Optional[Any] = 384 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Dict = 37 UpperCamelCase : Any = "gelu" UpperCamelCase : List[Any] = 0.1 UpperCamelCase : int = 0.1 UpperCamelCase : Tuple = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : int = 2 UpperCamelCase : Dict = 0.02 UpperCamelCase : Optional[Any] = 3 UpperCamelCase : List[Any] = 4 UpperCamelCase : Dict = 128 UpperCamelCase : Optional[Any] = 2 UpperCamelCase : Optional[int] = 9 UpperCamelCase : Optional[int] = 1 UpperCamelCase : Union[str, Any] = None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None if self.use_token_type_ids: UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel(config=A_ ) UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : Any = model(A_ ) UpperCamelCase : int = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = TFConvBertForMaskedLM(config=A_ ) UpperCamelCase : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = TFConvBertForSequenceClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : str = TFConvBertForMultipleChoice(config=A_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Dict = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Any = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : str = TFConvBertForTokenClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : str = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFConvBertForQuestionAnswering(config=A_ ) UpperCamelCase : Union[str, Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Union[str, Any] = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase :Any = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = TFConvBertModelTester(self ) UpperCamelCase : Dict = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Optional[Any] = True UpperCamelCase : Any = True if hasattr(A_ , "use_cache" ): UpperCamelCase : List[str] = True UpperCamelCase : List[Any] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Any = getattr(self.model_tester , "key_length" , A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[Any] = self._prepare_for_class(A_ , A_ ) UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Optional[int] = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) UpperCamelCase : Union[str, Any] = os.path.join(A_ , "saved_model" , "1" ) UpperCamelCase : Dict = tf.keras.models.load_model(A_ ) UpperCamelCase : str = model(A_ ) if self.is_encoder_decoder: UpperCamelCase : Union[str, Any] = outputs["encoder_hidden_states"] UpperCamelCase : Any = outputs["encoder_attentions"] else: UpperCamelCase : Any = outputs["hidden_states"] UpperCamelCase : List[str] = outputs["attentions"] self.assertEqual(len(A_ ) , A_ ) UpperCamelCase : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True UpperCamelCase : int = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "key_length" , A_ ) UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , A_ ) def check_decoder_attentions_output(A_ ): UpperCamelCase : Optional[Any] = len(A_ ) self.assertEqual(out_len % 2 , 0 ) UpperCamelCase : Any = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): UpperCamelCase : Dict = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True UpperCamelCase : List[Any] = False UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : List[str] = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Tuple = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase : Tuple = True UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine UpperCamelCase : Optional[int] = True UpperCamelCase : List[str] = True UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : Optional[Any] = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : List[str] = model(A_ )[0] UpperCamelCase : int = [1, 6, 768] self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
38
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __lowerCamelCase : Optional[int] = logging.get_logger(__name__) class A__ ( __snake_case , __snake_case ): _UpperCAmelCase :int = 'maskformer-swin' _UpperCAmelCase :Union[str, Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , A_=224 , A_=4 , A_=3 , A_=96 , A_=[2, 2, 6, 2] , A_=[3, 6, 12, 24] , A_=7 , A_=4.0 , A_=True , A_=0.0 , A_=0.0 , A_=0.1 , A_="gelu" , A_=False , A_=0.02 , A_=1e-5 , A_=None , A_=None , **A_ , ): '''simple docstring''' super().__init__(**A_ ) UpperCamelCase : Any = image_size UpperCamelCase : int = patch_size UpperCamelCase : Tuple = num_channels UpperCamelCase : str = embed_dim UpperCamelCase : Any = depths UpperCamelCase : str = len(A_ ) UpperCamelCase : List[str] = num_heads UpperCamelCase : str = window_size UpperCamelCase : str = mlp_ratio UpperCamelCase : Optional[int] = qkv_bias UpperCamelCase : Union[str, Any] = hidden_dropout_prob UpperCamelCase : Any = attention_probs_dropout_prob UpperCamelCase : Dict = drop_path_rate UpperCamelCase : List[Any] = hidden_act UpperCamelCase : Tuple = use_absolute_embeddings UpperCamelCase : Dict = layer_norm_eps UpperCamelCase : Optional[int] = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCamelCase : List[Any] = int(embed_dim * 2 ** (len(A_ ) - 1) ) UpperCamelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(A_ ) + 1 )] UpperCamelCase , UpperCamelCase : Optional[int] = get_aligned_output_features_output_indices( out_features=A_ , out_indices=A_ , stage_names=self.stage_names )
38
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : str = { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'camembert' def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : List[str] = vocab_size UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : Dict = hidden_act UpperCamelCase : str = intermediate_size UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : int = initializer_range UpperCamelCase : List[str] = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : int = use_cache UpperCamelCase : List[str] = classifier_dropout class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : str = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
38
1
import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : List[str] = False, False, False @dataclass class A__ : _UpperCAmelCase :Optional[int] = None _UpperCAmelCase :bool = True _UpperCAmelCase :bool = True _UpperCAmelCase :Optional[str] = None # Automatically constructed _UpperCAmelCase :ClassVar[str] = "dict" _UpperCAmelCase :ClassVar[Any] = pa.struct({'bytes': pa.binary(), 'path': pa.string()} ) _UpperCAmelCase :str = field(default='Audio' , init=__snake_case , repr=__snake_case ) def __call__( self ): '''simple docstring''' return self.pa_type def __UpperCamelCase( self , A_ ): '''simple docstring''' try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("To support encoding audio data, please install 'soundfile'." ) from err if isinstance(A_ , A_ ): return {"bytes": None, "path": value} elif isinstance(A_ , A_ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes UpperCamelCase : Optional[int] = BytesIO() sf.write(A_ , value["array"] , value["sampling_rate"] , format="wav" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("pcm" ): # "PCM" only has raw audio bytes if value.get("sampling_rate" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("To use PCM files, please specify a 'sampling_rate' in Audio object" ) if value.get("bytes" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) UpperCamelCase : Dict = np.frombuffer(value["bytes"] , dtype=np.intaa ).astype(np.floataa ) / 3_2767 else: UpperCamelCase : Any = np.memmap(value["path"] , dtype="h" , mode="r" ).astype(np.floataa ) / 3_2767 UpperCamelCase : List[str] = BytesIO(bytes() ) sf.write(A_ , A_ , value["sampling_rate"] , format="wav" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( F"""An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""" ) def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Audio(decode=True) instead." ) UpperCamelCase , UpperCamelCase : Optional[int] = (value["path"], BytesIO(value["bytes"] )) if value["bytes"] is not None else (value["path"], None) if path is None and file is None: raise ValueError(F"""An audio sample should have one of 'path' or 'bytes' but both are None in {value}.""" ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("To support decoding audio files, please install 'librosa' and 'soundfile'." ) from err UpperCamelCase : Optional[Any] = xsplitext(A_ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( "Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( "Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) if file is None: UpperCamelCase : Optional[Any] = token_per_repo_id or {} UpperCamelCase : List[Any] = path.split("::" )[-1] try: UpperCamelCase : str = string_to_dict(A_ , config.HUB_DATASETS_URL )["repo_id"] UpperCamelCase : Optional[Any] = token_per_repo_id[repo_id] except (ValueError, KeyError): UpperCamelCase : Union[str, Any] = None with xopen(A_ , "rb" , use_auth_token=A_ ) as f: UpperCamelCase , UpperCamelCase : List[str] = sf.read(A_ ) else: UpperCamelCase , UpperCamelCase : Any = sf.read(A_ ) UpperCamelCase : Tuple = array.T if self.mono: UpperCamelCase : Union[str, Any] = librosa.to_mono(A_ ) if self.sampling_rate and self.sampling_rate != sampling_rate: UpperCamelCase : Optional[Any] = librosa.resample(A_ , orig_sr=A_ , target_sr=self.sampling_rate ) UpperCamelCase : List[str] = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def __UpperCamelCase( self ): '''simple docstring''' from .features import Value if self.decode: raise ValueError("Cannot flatten a decoded Audio feature." ) return { "bytes": Value("binary" ), "path": Value("string" ), } def __UpperCamelCase( self , A_ ): '''simple docstring''' if pa.types.is_string(storage.type ): UpperCamelCase : Union[str, Any] = pa.array([None] * len(A_ ) , type=pa.binary() ) UpperCamelCase : Union[str, Any] = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCamelCase : str = pa.array([None] * len(A_ ) , type=pa.string() ) UpperCamelCase : Optional[int] = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("array" ): UpperCamelCase : Optional[int] = pa.array([Audio().encode_example(A_ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: UpperCamelCase : Tuple = storage.field("bytes" ) else: UpperCamelCase : List[Any] = pa.array([None] * len(A_ ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: UpperCamelCase : List[str] = storage.field("path" ) else: UpperCamelCase : Union[str, Any] = pa.array([None] * len(A_ ) , type=pa.string() ) UpperCamelCase : Dict = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) return array_cast(A_ , self.pa_type ) def __UpperCamelCase( self , A_ ): '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(A_ ): with xopen(A_ , "rb" ) as f: UpperCamelCase : int = f.read() return bytes_ UpperCamelCase : Union[str, Any] = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) UpperCamelCase : Optional[Any] = pa.array( [os.path.basename(A_ ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) UpperCamelCase : Tuple = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(A_ , self.pa_type )
38
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: return int(input_a == input_a == 0 ) def A_ ( ) -> None: print("Truth Table of NOR Gate:" ) print("| Input 1 | Input 2 | Output |" ) print(F"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> str: if config_name_or_path is None: UpperCamelCase : Dict = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCamelCase : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCamelCase : Tuple = question_encoder_name_or_path UpperCamelCase : Any = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCamelCase : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Tuple = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : int = gen_config UpperCamelCase : Dict = question_encoder_config UpperCamelCase : Tuple = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--model_type""", choices=["""rag_sequence""", """rag_token"""], required=True, type=str, help="""RAG model type: rag_sequence, rag_token""", ) parser.add_argument("""--dest""", type=str, required=True, help="""Path to the output checkpoint directory.""") parser.add_argument("""--generator_name_or_path""", type=str, required=True, help="""Generator model identifier""") parser.add_argument( """--question_encoder_name_or_path""", type=str, required=True, help="""Question encoder model identifier""" ) parser.add_argument( """--generator_tokenizer_name_or_path""", type=str, help="""Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``""", ) parser.add_argument( """--question_encoder_tokenizer_name_or_path""", type=str, help="""Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``""", ) parser.add_argument( """--config_name_or_path""", type=str, help=( """Identifier of the model config to use, if not provided, resolves to a base config for a given""" """ ``model_type``""" ), ) __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
38
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __snake_case ): _UpperCAmelCase :Optional[int] = ['image_processor', 'tokenizer'] _UpperCAmelCase :Tuple = 'BlipImageProcessor' _UpperCAmelCase :Optional[int] = 'AutoTokenizer' def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = False super().__init__(A_ , A_ ) UpperCamelCase : str = self.image_processor def __call__( self , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: UpperCamelCase : int = self.tokenizer UpperCamelCase : Optional[int] = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) return text_encoding # add pixel_values UpperCamelCase : int = self.image_processor(A_ , return_tensors=A_ ) if text is not None: UpperCamelCase : Dict = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) else: UpperCamelCase : Dict = None if text_encoding is not None: encoding_image_processor.update(A_ ) return encoding_image_processor def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.batch_decode(*A_ , **A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.decode(*A_ , **A_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.tokenizer.model_input_names UpperCamelCase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
38
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[int] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys __lowerCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __lowerCamelCase : Dict = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Tuple = ['audio_values', 'audio_mask'] def __init__( self , A_=2048 , A_=1 , A_=[16, 16] , A_=128 , A_=4_4100 , A_=86 , A_=2048 , A_=0.0 , **A_ , ): '''simple docstring''' super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , **A_ , ) UpperCamelCase : Optional[int] = spectrogram_length UpperCamelCase : Dict = num_channels UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : str = feature_size // self.patch_size[1] UpperCamelCase : List[str] = n_fft UpperCamelCase : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase : Optional[int] = sampling_rate UpperCamelCase : int = padding_value UpperCamelCase : str = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=A_ , norm="slaney" , mel_scale="slaney" , ).T def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = spectrogram( A_ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) UpperCamelCase : List[Any] = log_spec[:, :-1] UpperCamelCase : Optional[int] = log_spec - 20.0 UpperCamelCase : str = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A_ , A_ = None , A_ = True , A_ = None , A_ = False , A_ = False , **A_ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase : Optional[int] = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase : Union[str, Any] = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): UpperCamelCase : str = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase : str = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A_ ): UpperCamelCase : int = [np.asarray(A_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase : Tuple = np.array(A_ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase : Union[str, Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase : Any = np.ones([len(A_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase : List[str] = padded_audio_features * self.padding_value for i in range(len(A_ ) ): UpperCamelCase : Union[str, Any] = audio_features[i] UpperCamelCase : Optional[int] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase : Optional[Any] = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: UpperCamelCase : int = {"audio_values": padded_audio_features} UpperCamelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
38
1
def A_ ( ) -> Optional[Any]: UpperCamelCase : List[str] = [] UpperCamelCase : Any = 1 while len(_lowerCAmelCase ) < 1e6: constant.append(str(_lowerCAmelCase ) ) i += 1 UpperCamelCase : Any = "".join(_lowerCAmelCase ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[9_9999] ) * int(constant[99_9999] ) ) if __name__ == "__main__": print(solution())
38
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Dict = TypeVar("""KT""") __lowerCamelCase : Dict = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self , A_ = "root" , A_ = None ): '''simple docstring''' UpperCamelCase : int = key UpperCamelCase : List[Any] = value UpperCamelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F"""Node({self.key}: {self.value})""" @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self , A_ = 0.5 , A_ = 16 ): '''simple docstring''' UpperCamelCase : Node[KT, VT] = Node[KT, VT]() UpperCamelCase : List[Any] = 0 UpperCamelCase : Union[str, Any] = p UpperCamelCase : List[str] = max_level def __str__( self ): '''simple docstring''' UpperCamelCase : int = list(self ) if len(A_ ) == 0: return F"""SkipList(level={self.level})""" UpperCamelCase : str = max((len(str(A_ ) ) for item in items) , default=4 ) UpperCamelCase : Dict = max(A_ , 4 ) + 4 UpperCamelCase : str = self.head UpperCamelCase : List[Any] = [] UpperCamelCase : int = node.forward.copy() lines.append(F"""[{node.key}]""".ljust(A_ , "-" ) + "* " * len(A_ ) ) lines.append(" " * label_size + "| " * len(A_ ) ) while len(node.forward ) != 0: UpperCamelCase : Union[str, Any] = node.forward[0] lines.append( F"""[{node.key}]""".ljust(A_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(A_ ) ) UpperCamelCase : Tuple = node.forward lines.append("None".ljust(A_ ) + "* " * len(A_ ) ) return F"""SkipList(level={self.level})\n""" + "\n".join(A_ ) def __iter__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCamelCase : Union[str, Any] = node.forward[0] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = 1 while random() < self.p and level < self.max_level: level += 1 return level def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = [] UpperCamelCase : List[Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCamelCase : str = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(A_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self._locate_node(A_ ) if node is not None: for i, update_node in enumerate(A_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCamelCase : Tuple = node.forward[i] else: UpperCamelCase : List[Any] = update_node.forward[:i] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self._locate_node(A_ ) if node is not None: UpperCamelCase : Union[str, Any] = value else: UpperCamelCase : Dict = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , A_ ): update_vector.append(self.head ) UpperCamelCase : Optional[int] = level UpperCamelCase : Dict = Node(A_ , A_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(A_ ) else: UpperCamelCase : List[Any] = new_node def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self._locate_node(A_ ) if node is not None: return node.value return None def A_ ( ) -> List[Any]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) UpperCamelCase : Optional[int] = skip_list.head UpperCamelCase : List[str] = {} while node.level != 0: UpperCamelCase : str = node.forward[0] UpperCamelCase : Optional[int] = node.value assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def A_ ( ) -> List[Any]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) UpperCamelCase : Dict = skip_list.head UpperCamelCase : Tuple = {} while node.level != 0: UpperCamelCase : List[str] = node.forward[0] UpperCamelCase : Dict = node.value if len(_lowerCAmelCase ) != 4: print() assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() assert skip_list.find("Some key" ) is None def A_ ( ) -> Tuple: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def A_ ( ) -> Dict: UpperCamelCase : Optional[int] = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def A_ ( ) -> Dict: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[str]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_lowerCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_lowerCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def A_ ( ) -> Union[str, Any]: def is_sorted(_lowerCAmelCase ): return all(next_item >= item for item, next_item in zip(_lowerCAmelCase , lst[1:] ) ) UpperCamelCase : int = SkipList() for i in range(10 ): skip_list.insert(_lowerCAmelCase , _lowerCAmelCase ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_lowerCAmelCase ) ) def A_ ( ) -> Tuple: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def A_ ( ) -> List[str]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
from manim import * class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = Rectangle(height=0.5 , width=0.5 ) UpperCamelCase : str = Rectangle(height=0.25 , width=0.25 ) UpperCamelCase : Optional[Any] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCamelCase : Optional[Any] = [mem.copy() for i in range(6 )] UpperCamelCase : List[str] = [mem.copy() for i in range(6 )] UpperCamelCase : int = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Tuple = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Union[str, Any] = VGroup(A_ , A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Optional[int] = Text("CPU" , font_size=24 ) UpperCamelCase : List[Any] = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(A_ ) UpperCamelCase : List[Any] = [mem.copy() for i in range(4 )] UpperCamelCase : List[Any] = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Optional[Any] = Text("GPU" , font_size=24 ) UpperCamelCase : List[str] = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) gpu.move_to([-1, -1, 0] ) self.add(A_ ) UpperCamelCase : Any = [mem.copy() for i in range(6 )] UpperCamelCase : Any = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Optional[Any] = Text("Model" , font_size=24 ) UpperCamelCase : List[Any] = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) model.move_to([3, -1.0, 0] ) self.add(A_ ) UpperCamelCase : Dict = [] UpperCamelCase : int = [] UpperCamelCase : Tuple = [] for i, rect in enumerate(A_ ): rect.set_stroke(A_ ) UpperCamelCase : Optional[Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(A_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=A_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=A_ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=A_ , buff=0.0 ) self.add(A_ ) model_cpu_arr.append(A_ ) self.add(*A_ , *A_ , *A_ ) UpperCamelCase : Any = [mem.copy() for i in range(6 )] UpperCamelCase : List[str] = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Dict = Text("Loaded Checkpoint" , font_size=24 ) UpperCamelCase : Union[str, Any] = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) checkpoint.move_to([3, 0.5, 0] ) self.add(A_ ) UpperCamelCase : int = [] UpperCamelCase : List[Any] = [] for i, rect in enumerate(A_ ): UpperCamelCase : int = fill.copy().set_fill(A_ , opacity=0.7 ) target.move_to(A_ ) ckpt_arr.append(A_ ) UpperCamelCase : Optional[Any] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(A_ ) self.add(*A_ , *A_ ) UpperCamelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCamelCase : Any = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(A_ , A_ ) UpperCamelCase : List[str] = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(A_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(A_ ) UpperCamelCase : Optional[int] = MarkupText( F"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) UpperCamelCase : Union[str, Any] = [meta_mem.copy() for i in range(6 )] UpperCamelCase : List[str] = [meta_mem.copy() for i in range(6 )] UpperCamelCase : str = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : Union[str, Any] = VGroup(*A_ ).arrange(A_ , buff=0 ) UpperCamelCase : List[str] = VGroup(A_ , A_ ).arrange(A_ , buff=0 ) UpperCamelCase : str = Text("Disk" , font_size=24 ) UpperCamelCase : int = Group(A_ , A_ ).arrange(A_ , buff=0.5 , aligned_edge=A_ ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(A_ , run_time=3 ) , Write(A_ , run_time=1 ) , Create(A_ , run_time=1 ) ) UpperCamelCase : Union[str, Any] = [] for i, rect in enumerate(A_ ): UpperCamelCase : List[Any] = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(A_ , run_time=1.5 ) ) self.play(*A_ ) self.play(FadeOut(A_ ) ) UpperCamelCase : Any = MarkupText(F"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(A_ , run_time=3 ) ) self.play( FadeOut(A_ , A_ , *A_ , *A_ ) , ) self.wait()
38
from PIL import Image def A_ ( _lowerCAmelCase ) -> Image: UpperCamelCase , UpperCamelCase : List[Any] = image.size UpperCamelCase : Union[str, Any] = 0 UpperCamelCase : List[str] = image.load() for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(_lowerCAmelCase ): for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
38
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A_ , "tf_padding" ) ) self.parent.assertTrue(hasattr(A_ , "depth_multiplier" ) ) class A__ : def __init__( self , A_ , A_=13 , A_=3 , A_=32 , A_=0.25 , A_=8 , A_=True , A_=1024 , A_=32 , A_="relu6" , A_=0.1 , A_=0.02 , A_=True , A_=True , A_=10 , A_=None , ): '''simple docstring''' UpperCamelCase : Any = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : Optional[int] = num_channels UpperCamelCase : Tuple = image_size UpperCamelCase : Any = depth_multiplier UpperCamelCase : Any = min_depth UpperCamelCase : Dict = tf_padding UpperCamelCase : int = int(last_hidden_size * depth_multiplier ) UpperCamelCase : Optional[int] = output_stride UpperCamelCase : Any = hidden_act UpperCamelCase : Any = classifier_dropout_prob UpperCamelCase : Optional[int] = use_labels UpperCamelCase : List[str] = is_training UpperCamelCase : List[Any] = num_labels UpperCamelCase : Optional[Any] = initializer_range UpperCamelCase : Union[str, Any] = scope def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : int = None UpperCamelCase : Dict = None if self.use_labels: UpperCamelCase : Any = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase : Union[str, Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __UpperCamelCase( self ): '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = MobileNetVaModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : List[Any] = model(A_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = self.num_labels UpperCamelCase : List[str] = MobileNetVaForImageClassification(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Tuple = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : int = config_and_inputs UpperCamelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :List[Any] = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () _UpperCAmelCase :Optional[int] = ( {'feature-extraction': MobileNetVaModel, 'image-classification': MobileNetVaForImageClassification} if is_torch_available() else {} ) _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :List[str] = False _UpperCAmelCase :int = False _UpperCAmelCase :int = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = MobileNetVaModelTester(self ) UpperCamelCase : List[str] = MobileNetVaConfigTester(self , config_class=A_ , has_text_modality=A_ ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileNetV1 does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not support input and output embeddings" ) def __UpperCamelCase( self ): '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not output attentions" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : List[str] = model_class(A_ ) UpperCamelCase : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : List[Any] = [*signature.parameters.keys()] UpperCamelCase : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' def check_hidden_states_output(A_ , A_ , A_ ): UpperCamelCase : Dict = model_class(A_ ) model.to(A_ ) model.eval() with torch.no_grad(): UpperCamelCase : str = model(**self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : Union[str, Any] = outputs.hidden_states UpperCamelCase : List[str] = 26 self.assertEqual(len(A_ ) , A_ ) UpperCamelCase , UpperCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[int] = True check_hidden_states_output(A_ , A_ , A_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase : List[Any] = True check_hidden_states_output(A_ , A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Tuple = MobileNetVaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def A_ ( ) -> Any: UpperCamelCase : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def __UpperCamelCase( self ): '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained("google/mobilenet_v1_1.0_224" ) if is_vision_available() else None ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = MobileNetVaForImageClassification.from_pretrained("google/mobilenet_v1_1.0_224" ).to(A_ ) UpperCamelCase : List[Any] = self.default_image_processor UpperCamelCase : Dict = prepare_img() UpperCamelCase : Any = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Any = model(**A_ ) # verify the logits UpperCamelCase : Optional[int] = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , A_ ) UpperCamelCase : Any = torch.tensor([-4.17_39, -1.12_33, 3.12_05] ).to(A_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A_ , atol=1e-4 ) )
38
from math import loga def A_ ( _lowerCAmelCase ) -> int: if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
1
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
from __future__ import annotations __lowerCamelCase : Optional[int] = """Muhammad Umer Farooq""" __lowerCamelCase : Tuple = """MIT""" __lowerCamelCase : Optional[int] = """1.0.0""" __lowerCamelCase : int = """Muhammad Umer Farooq""" __lowerCamelCase : Optional[int] = """contact@muhammadumerfarooq.me""" __lowerCamelCase : Dict = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : list[str] = [] UpperCamelCase : str = domain def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCamelCase : Any = parse.urljoin(self.domain , A_ ) self.urls.append(A_ ) def A_ ( _lowerCAmelCase ) -> str: return ".".join(get_sub_domain_name(_lowerCAmelCase ).split("." )[-2:] ) def A_ ( _lowerCAmelCase ) -> str: return parse.urlparse(_lowerCAmelCase ).netloc def A_ ( _lowerCAmelCase = "https://github.com" ) -> list[str]: UpperCamelCase : int = get_domain_name(_lowerCAmelCase ) # Initialize the parser UpperCamelCase : str = Parser(_lowerCAmelCase ) try: # Open URL UpperCamelCase : int = requests.get(_lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCamelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCamelCase : Optional[Any] = requests.get(_lowerCAmelCase ) # Get the valid email. UpperCamelCase : Optional[int] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = emails_from_url("""https://github.com""") print(f"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
38
1
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : Tuple = { """deepmind/language-perceiver""": """https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json""", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class A__ ( __snake_case ): _UpperCAmelCase :int = 'perceiver' def __init__( self , A_=256 , A_=1280 , A_=768 , A_=1 , A_=26 , A_=8 , A_=8 , A_=None , A_=None , A_="kv" , A_=1 , A_=1 , A_="gelu" , A_=0.1 , A_=0.02 , A_=1e-12 , A_=True , A_=262 , A_=2048 , A_=56 , A_=[368, 496] , A_=16 , A_=1920 , A_=16 , A_=[1, 16, 224, 224] , **A_ , ): '''simple docstring''' super().__init__(**A_ ) UpperCamelCase : List[Any] = num_latents UpperCamelCase : int = d_latents UpperCamelCase : List[Any] = d_model UpperCamelCase : Union[str, Any] = num_blocks UpperCamelCase : List[Any] = num_self_attends_per_block UpperCamelCase : List[str] = num_self_attention_heads UpperCamelCase : str = num_cross_attention_heads UpperCamelCase : Optional[Any] = qk_channels UpperCamelCase : List[Any] = v_channels UpperCamelCase : Tuple = cross_attention_shape_for_attention UpperCamelCase : Optional[int] = self_attention_widening_factor UpperCamelCase : List[str] = cross_attention_widening_factor UpperCamelCase : List[str] = hidden_act UpperCamelCase : List[str] = attention_probs_dropout_prob UpperCamelCase : Optional[int] = initializer_range UpperCamelCase : Any = layer_norm_eps UpperCamelCase : Optional[Any] = use_query_residual # masked language modeling attributes UpperCamelCase : Optional[int] = vocab_size UpperCamelCase : Optional[int] = max_position_embeddings # image classification attributes UpperCamelCase : Dict = image_size # flow attributes UpperCamelCase : str = train_size # multimodal autoencoding attributes UpperCamelCase : Optional[int] = num_frames UpperCamelCase : List[Any] = audio_samples_per_frame UpperCamelCase : int = samples_per_patch UpperCamelCase : str = output_shape class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Any = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : Any = {0: "batch", 1: "sequence"} return OrderedDict( [ ("inputs", dynamic_axis), ("attention_mask", dynamic_axis), ] ) @property def __UpperCamelCase( self ): '''simple docstring''' return 1e-4 def __UpperCamelCase( self , A_ , A_ = -1 , A_ = -1 , A_ = -1 , A_ = False , A_ = None , A_ = 3 , A_ = 40 , A_ = 40 , ): '''simple docstring''' if isinstance(A_ , A_ ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCamelCase : Union[str, Any] = compute_effective_axis_dimension( A_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCamelCase : Optional[int] = preprocessor.num_special_tokens_to_add(A_ ) UpperCamelCase : List[Any] = compute_effective_axis_dimension( A_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A_ ) # Generate dummy inputs according to compute batch and sequence UpperCamelCase : str = [" ".join(["a"] ) * seq_length] * batch_size UpperCamelCase : Tuple = dict(preprocessor(A_ , return_tensors=A_ ) ) UpperCamelCase : int = inputs.pop("input_ids" ) return inputs elif isinstance(A_ , A_ ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCamelCase : List[str] = compute_effective_axis_dimension(A_ , fixed_dimension=OnnxConfig.default_fixed_batch ) UpperCamelCase : List[Any] = self._generate_dummy_images(A_ , A_ , A_ , A_ ) UpperCamelCase : Tuple = dict(preprocessor(images=A_ , return_tensors=A_ ) ) UpperCamelCase : Any = inputs.pop("pixel_values" ) return inputs else: raise ValueError( "Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor." )
38
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> list[int]: UpperCamelCase : Optional[Any] = [True] * limit UpperCamelCase : Optional[Any] = False UpperCamelCase : List[str] = False UpperCamelCase : Tuple = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCamelCase : Optional[Any] = i * 2 while index < limit: UpperCamelCase : int = False UpperCamelCase : Optional[int] = index + i UpperCamelCase : Any = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def A_ ( _lowerCAmelCase = 100_0000 ) -> int: UpperCamelCase : Union[str, Any] = prime_sieve(_lowerCAmelCase ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCamelCase : Dict = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCamelCase : int = j - i UpperCamelCase : Dict = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
38
1
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __snake_case ): def __init__( self , A_ , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , A_ = None , **A_ , ): '''simple docstring''' super().__init__( features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , num_proc=A_ , **A_ , ) UpperCamelCase : Optional[int] = Generator( cache_dir=A_ , features=A_ , generator=A_ , gen_kwargs=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' if self.streaming: UpperCamelCase : Optional[Any] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: UpperCamelCase : Union[str, Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , num_proc=self.num_proc , ) UpperCamelCase : int = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset
38
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[Any] = logging.get_logger(__name__) __lowerCamelCase : int = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'glpn' def __init__( self , A_=3 , A_=4 , A_=[2, 2, 2, 2] , A_=[8, 4, 2, 1] , A_=[32, 64, 160, 256] , A_=[7, 3, 3, 3] , A_=[4, 2, 2, 2] , A_=[1, 2, 5, 8] , A_=[4, 4, 4, 4] , A_="gelu" , A_=0.0 , A_=0.0 , A_=0.02 , A_=0.1 , A_=1e-6 , A_=64 , A_=10 , A_=-1 , **A_ , ): '''simple docstring''' super().__init__(**A_ ) UpperCamelCase : Tuple = num_channels UpperCamelCase : Tuple = num_encoder_blocks UpperCamelCase : Any = depths UpperCamelCase : Union[str, Any] = sr_ratios UpperCamelCase : Optional[Any] = hidden_sizes UpperCamelCase : List[str] = patch_sizes UpperCamelCase : str = strides UpperCamelCase : Tuple = mlp_ratios UpperCamelCase : str = num_attention_heads UpperCamelCase : Tuple = hidden_act UpperCamelCase : List[str] = hidden_dropout_prob UpperCamelCase : List[str] = attention_probs_dropout_prob UpperCamelCase : str = initializer_range UpperCamelCase : List[Any] = drop_path_rate UpperCamelCase : str = layer_norm_eps UpperCamelCase : Tuple = decoder_hidden_size UpperCamelCase : Tuple = max_depth UpperCamelCase : List[str] = head_in_index
38
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
1
def A_ ( _lowerCAmelCase ) -> bool: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise ValueError("check_bouncy() accepts only integer arguments" ) UpperCamelCase : Dict = str(_lowerCAmelCase ) UpperCamelCase : Tuple = "".join(sorted(_lowerCAmelCase ) ) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def A_ ( _lowerCAmelCase = 99 ) -> int: if not 0 < percent < 100: raise ValueError("solution() only accepts values from 0 to 100" ) UpperCamelCase : Any = 0 UpperCamelCase : Optional[int] = 1 while True: if check_bouncy(_lowerCAmelCase ): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(f"""{solution(99)}""")
38
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCamelCase : Dict = { """configuration_efficientnet""": [ """EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientNetConfig""", """EfficientNetOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Any = ["""EfficientNetImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[Any] = [ """EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientNetForImageClassification""", """EfficientNetModel""", """EfficientNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys __lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
38
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
1
# limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( """pipelines_utils""", """0.22.0""", """Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.""", standard_warn=False, stacklevel=3, )
38
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
1
import heapq as hq import math from collections.abc import Iterator class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = str(id_ ) UpperCamelCase : Dict = None UpperCamelCase : Tuple = None UpperCamelCase : Tuple = [] UpperCamelCase : List[str] = {} # {vertex:distance} def __lt__( self , A_ ): '''simple docstring''' return self.key < other.key def __repr__( self ): '''simple docstring''' return self.id def __UpperCamelCase( self , A_ ): '''simple docstring''' self.neighbors.append(A_ ) def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = weight def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , _lowerCAmelCase ) graph[b - 1].add_edge(graph[a - 1] , _lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> list: UpperCamelCase : int = [] for u in graph: UpperCamelCase : List[Any] = math.inf UpperCamelCase : Optional[Any] = None UpperCamelCase : Tuple = 0 UpperCamelCase : Any = graph[:] while q: UpperCamelCase : Any = min(_lowerCAmelCase ) q.remove(_lowerCAmelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCamelCase : Optional[int] = u UpperCamelCase : Tuple = u.edges[v.id] for i in range(1 , len(_lowerCAmelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Iterator[tuple]: for u in graph: UpperCamelCase : Any = math.inf UpperCamelCase : str = None UpperCamelCase : List[Any] = 0 UpperCamelCase : Tuple = list(_lowerCAmelCase ) hq.heapify(_lowerCAmelCase ) while h: UpperCamelCase : Any = hq.heappop(_lowerCAmelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCamelCase : Optional[Any] = u UpperCamelCase : Optional[int] = u.edges[v.id] hq.heapify(_lowerCAmelCase ) for i in range(1 , len(_lowerCAmelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def A_ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
38
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A_ ( ) -> Dict: UpperCamelCase : Tuple = ArgumentParser( description=( "PyTorch TPU distributed training launch " "helper utility that will spawn up " "multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCAmelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCAmelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCAmelCase ) return parser.parse_args() def A_ ( ) -> Optional[int]: UpperCamelCase : Tuple = parse_args() # Import training_script as a module. UpperCamelCase : Union[str, Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase : List[Any] = script_fpath.stem UpperCamelCase : Optional[Any] = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv UpperCamelCase : List[Any] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
38
1
from math import factorial def A_ ( _lowerCAmelCase = 100 ) -> int: return sum(int(_lowerCAmelCase ) for x in str(factorial(_lowerCAmelCase ) ) ) if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
38
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
1
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=64 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : List[str] = parent UpperCamelCase : Union[str, Any] = batch_size UpperCamelCase : Union[str, Any] = seq_length UpperCamelCase : List[Any] = is_training UpperCamelCase : Optional[int] = use_input_mask UpperCamelCase : List[Any] = use_token_type_ids UpperCamelCase : int = use_labels UpperCamelCase : str = vocab_size UpperCamelCase : List[Any] = hidden_size UpperCamelCase : List[Any] = embedding_size UpperCamelCase : Dict = num_hidden_layers UpperCamelCase : str = num_attention_heads UpperCamelCase : Union[str, Any] = intermediate_size UpperCamelCase : Dict = hidden_act UpperCamelCase : Dict = hidden_dropout_prob UpperCamelCase : int = attention_probs_dropout_prob UpperCamelCase : Optional[Any] = max_position_embeddings UpperCamelCase : Tuple = type_vocab_size UpperCamelCase : int = type_sequence_label_size UpperCamelCase : Any = initializer_range UpperCamelCase : int = num_labels UpperCamelCase : List[str] = num_choices UpperCamelCase : Union[str, Any] = scope def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : Dict = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : List[Any] = None if self.use_token_type_ids: UpperCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Tuple = None UpperCamelCase : int = None UpperCamelCase : Dict = None if self.use_labels: UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : str = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self ): '''simple docstring''' return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = MegatronBertModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Any = model(A_ , attention_mask=A_ , token_type_ids=A_ ) UpperCamelCase : Optional[Any] = model(A_ , token_type_ids=A_ ) UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = MegatronBertForMaskedLM(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Tuple = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = MegatronBertForCausalLM(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = MegatronBertForNextSentencePrediction(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Union[str, Any] = model( A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = MegatronBertForPreTraining(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : List[Any] = model( A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ , next_sentence_label=A_ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = MegatronBertForQuestionAnswering(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Union[str, Any] = model( A_ , attention_mask=A_ , token_type_ids=A_ , start_positions=A_ , end_positions=A_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = self.num_labels UpperCamelCase : Any = MegatronBertForSequenceClassification(A_ ) model.to(A_ ) model.eval() UpperCamelCase : List[str] = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = self.num_labels UpperCamelCase : int = MegatronBertForTokenClassification(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : str = model(A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : int = MegatronBertForMultipleChoice(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Optional[int] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Optional[Any] = model( A_ , attention_mask=A_ , token_type_ids=A_ , labels=A_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : str = config_and_inputs UpperCamelCase : Any = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :int = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) _UpperCAmelCase :int = ( { 'feature-extraction': MegatronBertModel, 'fill-mask': MegatronBertForMaskedLM, 'question-answering': MegatronBertForQuestionAnswering, 'text-classification': MegatronBertForSequenceClassification, 'text-generation': MegatronBertForCausalLM, 'token-classification': MegatronBertForTokenClassification, 'zero-shot': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase :List[Any] = True # test_resize_embeddings = False _UpperCAmelCase :Optional[Any] = False def __UpperCamelCase( self , A_ , A_ , A_=False ): '''simple docstring''' UpperCamelCase : Any = super()._prepare_for_class(A_ , A_ , return_labels=A_ ) if return_labels: if model_class in get_values(A_ ): UpperCamelCase : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=A_ ) UpperCamelCase : Dict = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A_ ) return inputs_dict def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = MegatronBertModelTester(self ) UpperCamelCase : str = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*A_ ) def A_ ( _lowerCAmelCase ) -> List[Any]: return torch.tensor( _lowerCAmelCase , dtype=torch.long , device=_lowerCAmelCase , ) __lowerCamelCase : int = 1E-4 @require_torch @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase ): @slow @unittest.skip("Model is not available." ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = "nvidia/megatron-bert-uncased-345m" if "MYDIR" in os.environ: UpperCamelCase : Any = os.path.join(os.environ["MYDIR"] , A_ ) UpperCamelCase : str = MegatronBertModel.from_pretrained(A_ ) model.to(A_ ) model.half() UpperCamelCase : Tuple = _long_tensor([[101, 7110, 1005, 1056, 2023, 1_1333, 1_7413, 1029, 102]] ) with torch.no_grad(): UpperCamelCase : str = model(A_ )[0] UpperCamelCase : Optional[Any] = torch.Size((1, 9, 1024) ) self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = [-0.60_40, -0.25_17, -0.10_25, 0.34_20, -0.67_58, -0.00_17, -0.10_89, -0.19_90, 0.57_28] for ii in range(3 ): for jj in range(3 ): UpperCamelCase : List[str] = output[0, ii, jj] UpperCamelCase : int = expected[3 * ii + jj] UpperCamelCase : int = "ii={} jj={} a={} b={}".format(A_ , A_ , A_ , A_ ) self.assertTrue(math.isclose(A_ , A_ , rel_tol=A_ , abs_tol=A_ ) , msg=A_ )
38
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __lowerCamelCase : Dict = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __lowerCamelCase : Tuple = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = VOCAB_FILES_NAMES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[Any] = BlenderbotSmallTokenizer def __init__( self , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , A_=True , **A_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=A_ , merges=A_ , add_prefix_space=A_ , trim_offsets=A_ , ) , bos_token=A_ , eos_token=A_ , unk_token=A_ , **A_ , ) UpperCamelCase : Union[str, Any] = add_prefix_space def __UpperCamelCase( self , A_ , A_=None ): '''simple docstring''' UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
38
1
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[Any] = KandinskyVaaInpaintPipeline _UpperCAmelCase :List[Any] = ['image_embeds', 'negative_image_embeds', 'image', 'mask_image'] _UpperCAmelCase :List[Any] = [ 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] _UpperCAmelCase :List[Any] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :Optional[int] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[Any] = { "in_channels": 9, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Optional[int] = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Union[str, Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Dict = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : int = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Any = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) # create init_image UpperCamelCase : Any = floats_tensor((1, 3, 64, 64) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCamelCase : str = Image.fromarray(np.uinta(A_ ) ).convert("RGB" ).resize((256, 256) ) # create mask UpperCamelCase : Dict = np.ones((64, 64) , dtype=np.floataa ) UpperCamelCase : Union[str, Any] = 0 if str(A_ ).startswith("mps" ): UpperCamelCase : Tuple = torch.manual_seed(A_ ) else: UpperCamelCase : Optional[int] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image": init_image, "mask_image": mask, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 2, "guidance_scale": 4.0, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = "cpu" UpperCamelCase : Dict = self.get_dummy_components() UpperCamelCase : str = self.pipeline_class(**A_ ) UpperCamelCase : Any = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : int = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : List[Any] = output.images UpperCamelCase : List[str] = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] UpperCamelCase : int = image_from_tuple[0, -3:, -3:, -1] print(F"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) UpperCamelCase : List[str] = np.array( [0.50_77_59_03, 0.49_52_71_95, 0.48_82_45_43, 0.50_19_22_37, 0.48_64_49_06, 0.49_37_38_14, 0.4_78_05_98, 0.47_23_48_27, 0.48_32_78_48] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def __UpperCamelCase( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_inpaint_cat_with_hat_fp16.npy" ) UpperCamelCase : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) UpperCamelCase : Tuple = np.ones((768, 768) , dtype=np.floataa ) UpperCamelCase : str = 0 UpperCamelCase : Tuple = "a hat" UpperCamelCase : Tuple = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Optional[Any] = KandinskyVaaInpaintPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder-inpaint" , torch_dtype=torch.floataa ) UpperCamelCase : int = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : List[Any] = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : Union[str, Any] = pipeline( image=A_ , mask_image=A_ , image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , height=768 , width=768 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(A_ , A_ )
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : int = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : Any = {"""configuration_xlnet""": ["""XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLNetConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Any = ["""XLNetTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Union[str, Any] = ["""XLNetTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[int] = [ """XLNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLNetForMultipleChoice""", """XLNetForQuestionAnswering""", """XLNetForQuestionAnsweringSimple""", """XLNetForSequenceClassification""", """XLNetForTokenClassification""", """XLNetLMHeadModel""", """XLNetModel""", """XLNetPreTrainedModel""", """load_tf_weights_in_xlnet""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXLNetForMultipleChoice""", """TFXLNetForQuestionAnsweringSimple""", """TFXLNetForSequenceClassification""", """TFXLNetForTokenClassification""", """TFXLNetLMHeadModel""", """TFXLNetMainLayer""", """TFXLNetModel""", """TFXLNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
import logging import os import threading import time try: import warnings except ImportError: __lowerCamelCase : str = None try: import msvcrt except ImportError: __lowerCamelCase : str = None try: import fcntl except ImportError: __lowerCamelCase : List[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: __lowerCamelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ __lowerCamelCase : str = [ """Timeout""", """BaseFileLock""", """WindowsFileLock""", """UnixFileLock""", """SoftFileLock""", """FileLock""", ] __lowerCamelCase : Union[str, Any] = """3.0.12""" __lowerCamelCase : Any = None def A_ ( ) -> List[Any]: global _logger UpperCamelCase : Any = _logger or logging.getLogger(__name__ ) return _logger class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = lock_file return None def __str__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = lock return None def __enter__( self ): '''simple docstring''' return self.lock def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.lock.release() return None class A__ : def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : List[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long UpperCamelCase : Dict = self.hash_filename_if_too_long(A_ , A_ ) # The path to the lock file. UpperCamelCase : List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. UpperCamelCase : Tuple = None # The default timeout value. UpperCamelCase : Optional[Any] = timeout # We use this lock primarily for the lock counter. UpperCamelCase : Union[str, Any] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. UpperCamelCase : Dict = 0 return None @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file @property def __UpperCamelCase( self ): '''simple docstring''' return self._timeout @timeout.setter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = float(A_ ) return None def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() def __UpperCamelCase( self ): '''simple docstring''' raise NotImplementedError() @property def __UpperCamelCase( self ): '''simple docstring''' return self._lock_file_fd is not None def __UpperCamelCase( self , A_=None , A_=0.05 ): '''simple docstring''' if timeout is None: UpperCamelCase : Optional[Any] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 UpperCamelCase : Dict = id(self ) UpperCamelCase : List[str] = self._lock_file UpperCamelCase : int = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(A_ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: UpperCamelCase : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __UpperCamelCase( self , A_=False ): '''simple docstring''' with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: UpperCamelCase : List[Any] = id(self ) UpperCamelCase : Dict = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() UpperCamelCase : Dict = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): '''simple docstring''' self.acquire() return self def __exit__( self , A_ , A_ , A_ ): '''simple docstring''' self.release() return None def __del__( self ): '''simple docstring''' self.release(force=A_ ) return None def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = os.path.basename(A_ ) if len(A_ ) > max_length and max_length > 0: UpperCamelCase : Optional[int] = os.path.dirname(A_ ) UpperCamelCase : int = str(hash(A_ ) ) UpperCamelCase : Any = filename[: max_length - len(A_ ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(A_ , A_ ) else: return path class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' from .file_utils import relative_to_absolute_path super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) UpperCamelCase : List[Any] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: UpperCamelCase : str = os.open(self._lock_file , A_ ) except OSError: pass else: try: msvcrt.locking(A_ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(A_ ) else: UpperCamelCase : Optional[Any] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self._lock_file_fd UpperCamelCase : str = None msvcrt.locking(A_ , msvcrt.LK_UNLCK , 1 ) os.close(A_ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class A__ ( __snake_case ): def __init__( self , A_ , A_=-1 , A_=None ): '''simple docstring''' UpperCamelCase : Tuple = os.statvfs(os.path.dirname(A_ ) ).f_namemax super().__init__(A_ , timeout=A_ , max_filename_length=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC UpperCamelCase : int = os.open(self._lock_file , A_ ) try: fcntl.flock(A_ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(A_ ) else: UpperCamelCase : List[str] = fd return None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self._lock_file_fd UpperCamelCase : List[Any] = None fcntl.flock(A_ , fcntl.LOCK_UN ) os.close(A_ ) return None class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: UpperCamelCase : Optional[int] = os.open(self._lock_file , A_ ) except OSError: pass else: UpperCamelCase : Tuple = fd return None def __UpperCamelCase( self ): '''simple docstring''' os.close(self._lock_file_fd ) UpperCamelCase : str = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None __lowerCamelCase : Dict = None if msvcrt: __lowerCamelCase : Any = WindowsFileLock elif fcntl: __lowerCamelCase : Any = UnixFileLock else: __lowerCamelCase : int = SoftFileLock if warnings is not None: warnings.warn("""only soft file lock is available""")
38
1
from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class A__ ( __snake_case ): def __init__( self , A_ , A_=None , A_=None , A_=0 ): '''simple docstring''' UpperCamelCase : Tuple = 1.0 if scale is None else scale UpperCamelCase : int = 0.0 if loc is None else loc super().__init__(A_ , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=A_ )] ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.base_dist.mean * self.scale + self.loc @property def __UpperCamelCase( self ): '''simple docstring''' return self.base_dist.variance * self.scale**2 @property def __UpperCamelCase( self ): '''simple docstring''' return self.variance.sqrt() class A__ ( nn.Module ): def __init__( self , A_ , A_ , A_ , **A_ ): '''simple docstring''' super().__init__(**A_ ) UpperCamelCase : Any = args_dim UpperCamelCase : List[Any] = nn.ModuleList([nn.Linear(A_ , A_ ) for dim in args_dim.values()] ) UpperCamelCase : Tuple = domain_map def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Dict = [proj(A_ ) for proj in self.proj] return self.domain_map(*A_ ) class A__ ( nn.Module ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : int = function def __UpperCamelCase( self , A_ , *A_ ): '''simple docstring''' return self.function(A_ , *A_ ) class A__ : _UpperCAmelCase :type _UpperCAmelCase :int _UpperCAmelCase :Dict[str, int] def __init__( self , A_ = 1 ): '''simple docstring''' UpperCamelCase : Tuple = dim UpperCamelCase : Optional[int] = {k: dim * self.args_dim[k] for k in self.args_dim} def __UpperCamelCase( self , A_ ): '''simple docstring''' if self.dim == 1: return self.distribution_class(*A_ ) else: return Independent(self.distribution_class(*A_ ) , 1 ) def __UpperCamelCase( self , A_ , A_ = None , A_ = None , ): '''simple docstring''' UpperCamelCase : List[Any] = self._base_distribution(A_ ) if loc is None and scale is None: return distr else: return AffineTransformed(A_ , loc=A_ , scale=A_ , event_dim=self.event_dim ) @property def __UpperCamelCase( self ): '''simple docstring''' return () if self.dim == 1 else (self.dim,) @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.event_shape ) @property def __UpperCamelCase( self ): '''simple docstring''' return 0.0 def __UpperCamelCase( self , A_ ): '''simple docstring''' return ParameterProjection( in_features=A_ , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def __UpperCamelCase( self , *A_ ): '''simple docstring''' raise NotImplementedError() @staticmethod def __UpperCamelCase( A_ ): '''simple docstring''' return (x + torch.sqrt(torch.square(A_ ) + 4.0 )) / 2.0 class A__ ( __snake_case ): _UpperCAmelCase :Dict[str, int] = {"df": 1, "loc": 1, "scale": 1} _UpperCAmelCase :type = StudentT @classmethod def __UpperCamelCase( cls , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Any = cls.squareplus(A_ ).clamp_min(torch.finfo(scale.dtype ).eps ) UpperCamelCase : Dict = 2.0 + cls.squareplus(A_ ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class A__ ( __snake_case ): _UpperCAmelCase :Dict[str, int] = {"loc": 1, "scale": 1} _UpperCAmelCase :type = Normal @classmethod def __UpperCamelCase( cls , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = cls.squareplus(A_ ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class A__ ( __snake_case ): _UpperCAmelCase :Dict[str, int] = {"total_count": 1, "logits": 1} _UpperCAmelCase :type = NegativeBinomial @classmethod def __UpperCamelCase( cls , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = cls.squareplus(A_ ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Any = distr_args if self.dim == 1: return self.distribution_class(total_count=A_ , logits=A_ ) else: return Independent(self.distribution_class(total_count=A_ , logits=A_ ) , 1 ) def __UpperCamelCase( self , A_ , A_ = None , A_ = None ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
38
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> str: if config_name_or_path is None: UpperCamelCase : Dict = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCamelCase : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCamelCase : Tuple = question_encoder_name_or_path UpperCamelCase : Any = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCamelCase : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : Tuple = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCamelCase : int = gen_config UpperCamelCase : Dict = question_encoder_config UpperCamelCase : Tuple = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--model_type""", choices=["""rag_sequence""", """rag_token"""], required=True, type=str, help="""RAG model type: rag_sequence, rag_token""", ) parser.add_argument("""--dest""", type=str, required=True, help="""Path to the output checkpoint directory.""") parser.add_argument("""--generator_name_or_path""", type=str, required=True, help="""Generator model identifier""") parser.add_argument( """--question_encoder_name_or_path""", type=str, required=True, help="""Question encoder model identifier""" ) parser.add_argument( """--generator_tokenizer_name_or_path""", type=str, help="""Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``""", ) parser.add_argument( """--question_encoder_tokenizer_name_or_path""", type=str, help="""Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``""", ) parser.add_argument( """--config_name_or_path""", type=str, help=( """Identifier of the model config to use, if not provided, resolves to a base config for a given""" """ ``model_type``""" ), ) __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
38
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __lowerCamelCase : Optional[int] = { """configuration_rag""": ["""RagConfig"""], """retrieval_rag""": ["""RagRetriever"""], """tokenization_rag""": ["""RagTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[Any] = [ """RagModel""", """RagPreTrainedModel""", """RagSequenceForGeneration""", """RagTokenForGeneration""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = [ """TFRagModel""", """TFRagPreTrainedModel""", """TFRagSequenceForGeneration""", """TFRagTokenForGeneration""", ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys __lowerCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : str = 13 UpperCamelCase : int = 7 UpperCamelCase : str = True UpperCamelCase : Dict = True UpperCamelCase : str = True UpperCamelCase : Tuple = True UpperCamelCase : List[str] = 99 UpperCamelCase : Optional[Any] = 384 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Dict = 37 UpperCamelCase : Any = "gelu" UpperCamelCase : List[Any] = 0.1 UpperCamelCase : int = 0.1 UpperCamelCase : Tuple = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : int = 2 UpperCamelCase : Dict = 0.02 UpperCamelCase : Optional[Any] = 3 UpperCamelCase : List[Any] = 4 UpperCamelCase : Dict = 128 UpperCamelCase : Optional[Any] = 2 UpperCamelCase : Optional[int] = 9 UpperCamelCase : Optional[int] = 1 UpperCamelCase : Union[str, Any] = None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None if self.use_token_type_ids: UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel(config=A_ ) UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : Any = model(A_ ) UpperCamelCase : int = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = TFConvBertForMaskedLM(config=A_ ) UpperCamelCase : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = TFConvBertForSequenceClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : str = TFConvBertForMultipleChoice(config=A_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Dict = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Any = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : str = TFConvBertForTokenClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : str = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFConvBertForQuestionAnswering(config=A_ ) UpperCamelCase : Union[str, Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Union[str, Any] = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase :Any = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = TFConvBertModelTester(self ) UpperCamelCase : Dict = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Optional[Any] = True UpperCamelCase : Any = True if hasattr(A_ , "use_cache" ): UpperCamelCase : List[str] = True UpperCamelCase : List[Any] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Any = getattr(self.model_tester , "key_length" , A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[Any] = self._prepare_for_class(A_ , A_ ) UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Optional[int] = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) UpperCamelCase : Union[str, Any] = os.path.join(A_ , "saved_model" , "1" ) UpperCamelCase : Dict = tf.keras.models.load_model(A_ ) UpperCamelCase : str = model(A_ ) if self.is_encoder_decoder: UpperCamelCase : Union[str, Any] = outputs["encoder_hidden_states"] UpperCamelCase : Any = outputs["encoder_attentions"] else: UpperCamelCase : Any = outputs["hidden_states"] UpperCamelCase : List[str] = outputs["attentions"] self.assertEqual(len(A_ ) , A_ ) UpperCamelCase : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True UpperCamelCase : int = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "key_length" , A_ ) UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , A_ ) def check_decoder_attentions_output(A_ ): UpperCamelCase : Optional[Any] = len(A_ ) self.assertEqual(out_len % 2 , 0 ) UpperCamelCase : Any = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): UpperCamelCase : Dict = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True UpperCamelCase : List[Any] = False UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : List[str] = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Tuple = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase : Tuple = True UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine UpperCamelCase : Optional[int] = True UpperCamelCase : List[str] = True UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : Optional[Any] = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : List[str] = model(A_ )[0] UpperCamelCase : int = [1, 6, 768] self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
38
1
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class A__ ( nn.Module ): def __init__( self ): '''simple docstring''' super().__init__() UpperCamelCase : Optional[Any] = nn.Linear(3 , 4 ) UpperCamelCase : Dict = nn.BatchNormad(4 ) UpperCamelCase : List[str] = nn.Linear(4 , 5 ) def __UpperCamelCase( self , A_ ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(A_ ) ) ) class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(A_ , model.state_dict() ) UpperCamelCase : Optional[Any] = os.path.join(A_ , "index.json" ) self.assertTrue(os.path.isfile(A_ ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: UpperCamelCase : Optional[int] = os.path.join(A_ , F"""{key}.dat""" ) self.assertTrue(os.path.isfile(A_ ) ) # TODO: add tests on the fact weights are properly loaded def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: UpperCamelCase : Any = torch.randn(2 , 3 , dtype=A_ ) with TemporaryDirectory() as tmp_dir: UpperCamelCase : str = offload_weight(A_ , "weight" , A_ , {} ) UpperCamelCase : Optional[Any] = os.path.join(A_ , "weight.dat" ) self.assertTrue(os.path.isfile(A_ ) ) self.assertDictEqual(A_ , {"weight": {"shape": [2, 3], "dtype": str(A_ ).split("." )[1]}} ) UpperCamelCase : List[Any] = load_offloaded_weight(A_ , index["weight"] ) self.assertTrue(torch.equal(A_ , A_ ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = ModelForTest() UpperCamelCase : Optional[int] = model.state_dict() UpperCamelCase : List[Any] = {k: v for k, v in state_dict.items() if "linear2" not in k} UpperCamelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(A_ , A_ ) UpperCamelCase : Optional[int] = OffloadedWeightsLoader(state_dict=A_ , save_folder=A_ ) # Every key is there with the right value self.assertEqual(sorted(A_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(A_ , weight_map[key] ) ) UpperCamelCase : str = {k: v for k, v in state_dict.items() if "weight" in k} UpperCamelCase : List[str] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(A_ , A_ ) UpperCamelCase : Any = OffloadedWeightsLoader(state_dict=A_ , save_folder=A_ ) # Every key is there with the right value self.assertEqual(sorted(A_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(A_ , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(A_ , A_ ) # Duplicates are removed UpperCamelCase : Dict = OffloadedWeightsLoader(state_dict=A_ , save_folder=A_ ) # Every key is there with the right value self.assertEqual(sorted(A_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(A_ , weight_map[key] ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = {"a.1": 0, "a.10": 1, "a.2": 2} UpperCamelCase : int = extract_submodules_state_dict(A_ , ["a.1", "a.2"] ) self.assertDictEqual(A_ , {"a.1": 0, "a.2": 2} ) UpperCamelCase : Tuple = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} UpperCamelCase : Dict = extract_submodules_state_dict(A_ , ["a.1", "a.2"] ) self.assertDictEqual(A_ , {"a.1.a": 0, "a.2.a": 2} )
38
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : str = { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'camembert' def __init__( self , A_=3_0522 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : List[str] = vocab_size UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : Any = num_hidden_layers UpperCamelCase : Union[str, Any] = num_attention_heads UpperCamelCase : Dict = hidden_act UpperCamelCase : str = intermediate_size UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : int = initializer_range UpperCamelCase : List[str] = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : int = use_cache UpperCamelCase : List[str] = classifier_dropout class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : str = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
38
1
def A_ ( _lowerCAmelCase ) -> int: if n == 1 or not isinstance(_lowerCAmelCase , _lowerCAmelCase ): return 0 elif n == 2: return 1 else: UpperCamelCase : Tuple = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def A_ ( _lowerCAmelCase ) -> int: UpperCamelCase : Optional[Any] = 0 UpperCamelCase : Optional[Any] = 2 while digits < n: index += 1 UpperCamelCase : Optional[int] = len(str(fibonacci(_lowerCAmelCase ) ) ) return index def A_ ( _lowerCAmelCase = 1000 ) -> int: return fibonacci_digits_index(_lowerCAmelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
38
def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: return int(input_a == input_a == 0 ) def A_ ( ) -> None: print("Truth Table of NOR Gate:" ) print("| Input 1 | Input 2 | Output |" ) print(F"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __lowerCamelCase : Dict = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Tuple = ['audio_values', 'audio_mask'] def __init__( self , A_=2048 , A_=1 , A_=[16, 16] , A_=128 , A_=4_4100 , A_=86 , A_=2048 , A_=0.0 , **A_ , ): '''simple docstring''' super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , **A_ , ) UpperCamelCase : Optional[int] = spectrogram_length UpperCamelCase : Dict = num_channels UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : str = feature_size // self.patch_size[1] UpperCamelCase : List[str] = n_fft UpperCamelCase : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase : Optional[int] = sampling_rate UpperCamelCase : int = padding_value UpperCamelCase : str = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=A_ , norm="slaney" , mel_scale="slaney" , ).T def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = spectrogram( A_ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) UpperCamelCase : List[Any] = log_spec[:, :-1] UpperCamelCase : Optional[int] = log_spec - 20.0 UpperCamelCase : str = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A_ , A_ = None , A_ = True , A_ = None , A_ = False , A_ = False , **A_ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase : Optional[int] = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase : Union[str, Any] = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): UpperCamelCase : str = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase : str = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A_ ): UpperCamelCase : int = [np.asarray(A_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase : Tuple = np.array(A_ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase : Union[str, Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase : Any = np.ones([len(A_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase : List[str] = padded_audio_features * self.padding_value for i in range(len(A_ ) ): UpperCamelCase : Union[str, Any] = audio_features[i] UpperCamelCase : Optional[int] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase : Optional[Any] = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: UpperCamelCase : int = {"audio_values": padded_audio_features} UpperCamelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
38
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __snake_case ): _UpperCAmelCase :Optional[int] = ['image_processor', 'tokenizer'] _UpperCAmelCase :Tuple = 'BlipImageProcessor' _UpperCAmelCase :Optional[int] = 'AutoTokenizer' def __init__( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = False super().__init__(A_ , A_ ) UpperCamelCase : str = self.image_processor def __call__( self , A_ = None , A_ = None , A_ = True , A_ = False , A_ = None , A_ = None , A_ = 0 , A_ = None , A_ = None , A_ = False , A_ = False , A_ = False , A_ = False , A_ = False , A_ = True , A_ = None , **A_ , ): '''simple docstring''' if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: UpperCamelCase : int = self.tokenizer UpperCamelCase : Optional[int] = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) return text_encoding # add pixel_values UpperCamelCase : int = self.image_processor(A_ , return_tensors=A_ ) if text is not None: UpperCamelCase : Dict = self.tokenizer( text=A_ , add_special_tokens=A_ , padding=A_ , truncation=A_ , max_length=A_ , stride=A_ , pad_to_multiple_of=A_ , return_attention_mask=A_ , return_overflowing_tokens=A_ , return_special_tokens_mask=A_ , return_offsets_mapping=A_ , return_token_type_ids=A_ , return_length=A_ , verbose=A_ , return_tensors=A_ , **A_ , ) else: UpperCamelCase : Dict = None if text_encoding is not None: encoding_image_processor.update(A_ ) return encoding_image_processor def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.batch_decode(*A_ , **A_ ) def __UpperCamelCase( self , *A_ , **A_ ): '''simple docstring''' return self.tokenizer.decode(*A_ , **A_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.tokenizer.model_input_names UpperCamelCase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
38
1
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Dict = TypeVar("""KT""") __lowerCamelCase : Dict = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self , A_ = "root" , A_ = None ): '''simple docstring''' UpperCamelCase : int = key UpperCamelCase : List[Any] = value UpperCamelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F"""Node({self.key}: {self.value})""" @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self , A_ = 0.5 , A_ = 16 ): '''simple docstring''' UpperCamelCase : Node[KT, VT] = Node[KT, VT]() UpperCamelCase : List[Any] = 0 UpperCamelCase : Union[str, Any] = p UpperCamelCase : List[str] = max_level def __str__( self ): '''simple docstring''' UpperCamelCase : int = list(self ) if len(A_ ) == 0: return F"""SkipList(level={self.level})""" UpperCamelCase : str = max((len(str(A_ ) ) for item in items) , default=4 ) UpperCamelCase : Dict = max(A_ , 4 ) + 4 UpperCamelCase : str = self.head UpperCamelCase : List[Any] = [] UpperCamelCase : int = node.forward.copy() lines.append(F"""[{node.key}]""".ljust(A_ , "-" ) + "* " * len(A_ ) ) lines.append(" " * label_size + "| " * len(A_ ) ) while len(node.forward ) != 0: UpperCamelCase : Union[str, Any] = node.forward[0] lines.append( F"""[{node.key}]""".ljust(A_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(A_ ) ) UpperCamelCase : Tuple = node.forward lines.append("None".ljust(A_ ) + "* " * len(A_ ) ) return F"""SkipList(level={self.level})\n""" + "\n".join(A_ ) def __iter__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCamelCase : Union[str, Any] = node.forward[0] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = 1 while random() < self.p and level < self.max_level: level += 1 return level def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = [] UpperCamelCase : List[Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCamelCase : str = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(A_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self._locate_node(A_ ) if node is not None: for i, update_node in enumerate(A_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCamelCase : Tuple = node.forward[i] else: UpperCamelCase : List[Any] = update_node.forward[:i] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self._locate_node(A_ ) if node is not None: UpperCamelCase : Union[str, Any] = value else: UpperCamelCase : Dict = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , A_ ): update_vector.append(self.head ) UpperCamelCase : Optional[int] = level UpperCamelCase : Dict = Node(A_ , A_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(A_ ) else: UpperCamelCase : List[Any] = new_node def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self._locate_node(A_ ) if node is not None: return node.value return None def A_ ( ) -> List[Any]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) UpperCamelCase : Optional[int] = skip_list.head UpperCamelCase : List[str] = {} while node.level != 0: UpperCamelCase : str = node.forward[0] UpperCamelCase : Optional[int] = node.value assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def A_ ( ) -> List[Any]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) UpperCamelCase : Dict = skip_list.head UpperCamelCase : Tuple = {} while node.level != 0: UpperCamelCase : List[str] = node.forward[0] UpperCamelCase : Dict = node.value if len(_lowerCAmelCase ) != 4: print() assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() assert skip_list.find("Some key" ) is None def A_ ( ) -> Tuple: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def A_ ( ) -> Dict: UpperCamelCase : Optional[int] = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def A_ ( ) -> Dict: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[str]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_lowerCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_lowerCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def A_ ( ) -> Union[str, Any]: def is_sorted(_lowerCAmelCase ): return all(next_item >= item for item, next_item in zip(_lowerCAmelCase , lst[1:] ) ) UpperCamelCase : int = SkipList() for i in range(10 ): skip_list.insert(_lowerCAmelCase , _lowerCAmelCase ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_lowerCAmelCase ) ) def A_ ( ) -> Tuple: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def A_ ( ) -> List[str]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __lowerCamelCase : Dict = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Tuple = ['audio_values', 'audio_mask'] def __init__( self , A_=2048 , A_=1 , A_=[16, 16] , A_=128 , A_=4_4100 , A_=86 , A_=2048 , A_=0.0 , **A_ , ): '''simple docstring''' super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , **A_ , ) UpperCamelCase : Optional[int] = spectrogram_length UpperCamelCase : Dict = num_channels UpperCamelCase : Optional[Any] = patch_size UpperCamelCase : str = feature_size // self.patch_size[1] UpperCamelCase : List[str] = n_fft UpperCamelCase : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase : Optional[int] = sampling_rate UpperCamelCase : int = padding_value UpperCamelCase : str = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=2_20_50.0 , sampling_rate=A_ , norm="slaney" , mel_scale="slaney" , ).T def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = spectrogram( A_ , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=80.0 , ) UpperCamelCase : List[Any] = log_spec[:, :-1] UpperCamelCase : Optional[int] = log_spec - 20.0 UpperCamelCase : str = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , A_ , A_ = None , A_ = True , A_ = None , A_ = False , A_ = False , **A_ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" F""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" F""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase : Optional[int] = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase : Union[str, Any] = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): UpperCamelCase : str = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase : str = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , A_ ): UpperCamelCase : int = [np.asarray(A_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase : str = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase : Tuple = np.array(A_ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase : Union[str, Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase : Any = np.ones([len(A_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase : List[str] = padded_audio_features * self.padding_value for i in range(len(A_ ) ): UpperCamelCase : Union[str, Any] = audio_features[i] UpperCamelCase : Optional[int] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase : Optional[Any] = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: UpperCamelCase : int = {"audio_values": padded_audio_features} UpperCamelCase : Any = BatchFeature(data=A_ , tensor_type=A_ ) return encoded_inputs
38
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import FEATURE_EXTRACTOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor, ChineseCLIPProcessor @require_vision class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = tempfile.mkdtemp() UpperCamelCase : int = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "的", "价", "格", "是", "15", "便", "alex", "##andra", ",", "。", "-", "t", "shirt", ] UpperCamelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) UpperCamelCase : Tuple = { "do_resize": True, "size": {"height": 224, "width": 224}, "do_center_crop": True, "crop_size": {"height": 18, "width": 18}, "do_normalize": True, "image_mean": [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], "image_std": [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], "do_convert_rgb": True, } UpperCamelCase : int = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(A_ , A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return BertTokenizer.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return BertTokenizerFast.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self , **A_ ): '''simple docstring''' return ChineseCLIPImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCamelCase : List[str] = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.get_tokenizer() UpperCamelCase : List[str] = self.get_rust_tokenizer() UpperCamelCase : List[Any] = self.get_image_processor() UpperCamelCase : Tuple = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase : Any = ChineseCLIPProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) UpperCamelCase : Optional[int] = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase : Union[str, Any] = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , A_ ) self.assertIsInstance(processor_fast.tokenizer , A_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , A_ ) self.assertIsInstance(processor_fast.image_processor , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = ChineseCLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase : Any = self.get_tokenizer(cls_token="(CLS)" , sep_token="(SEP)" ) UpperCamelCase : List[str] = self.get_image_processor(do_normalize=A_ ) UpperCamelCase : Union[str, Any] = ChineseCLIPProcessor.from_pretrained( self.tmpdirname , cls_token="(CLS)" , sep_token="(SEP)" , do_normalize=A_ ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.get_image_processor() UpperCamelCase : Tuple = self.get_tokenizer() UpperCamelCase : Union[str, Any] = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Optional[int] = self.prepare_image_inputs() UpperCamelCase : Tuple = image_processor(A_ , return_tensors="np" ) UpperCamelCase : List[str] = processor(images=A_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = self.get_image_processor() UpperCamelCase : Dict = self.get_tokenizer() UpperCamelCase : Tuple = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Any = "Alexandra,T-shirt的价格是15便士。" UpperCamelCase : List[str] = processor(text=A_ ) UpperCamelCase : Optional[int] = tokenizer(A_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.get_image_processor() UpperCamelCase : Optional[int] = self.get_tokenizer() UpperCamelCase : Optional[int] = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : List[Any] = "Alexandra,T-shirt的价格是15便士。" UpperCamelCase : Union[str, Any] = self.prepare_image_inputs() UpperCamelCase : List[str] = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "token_type_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.get_image_processor() UpperCamelCase : Optional[int] = self.get_tokenizer() UpperCamelCase : str = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase : Any = processor.batch_decode(A_ ) UpperCamelCase : Optional[int] = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.get_image_processor() UpperCamelCase : int = self.get_tokenizer() UpperCamelCase : List[str] = ChineseCLIPProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase : Any = "Alexandra,T-shirt的价格是15便士。" UpperCamelCase : Tuple = self.prepare_image_inputs() UpperCamelCase : int = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
38
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Dict = TypeVar("""KT""") __lowerCamelCase : Dict = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self , A_ = "root" , A_ = None ): '''simple docstring''' UpperCamelCase : int = key UpperCamelCase : List[Any] = value UpperCamelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F"""Node({self.key}: {self.value})""" @property def __UpperCamelCase( self ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self , A_ = 0.5 , A_ = 16 ): '''simple docstring''' UpperCamelCase : Node[KT, VT] = Node[KT, VT]() UpperCamelCase : List[Any] = 0 UpperCamelCase : Union[str, Any] = p UpperCamelCase : List[str] = max_level def __str__( self ): '''simple docstring''' UpperCamelCase : int = list(self ) if len(A_ ) == 0: return F"""SkipList(level={self.level})""" UpperCamelCase : str = max((len(str(A_ ) ) for item in items) , default=4 ) UpperCamelCase : Dict = max(A_ , 4 ) + 4 UpperCamelCase : str = self.head UpperCamelCase : List[Any] = [] UpperCamelCase : int = node.forward.copy() lines.append(F"""[{node.key}]""".ljust(A_ , "-" ) + "* " * len(A_ ) ) lines.append(" " * label_size + "| " * len(A_ ) ) while len(node.forward ) != 0: UpperCamelCase : Union[str, Any] = node.forward[0] lines.append( F"""[{node.key}]""".ljust(A_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(A_ ) ) UpperCamelCase : Tuple = node.forward lines.append("None".ljust(A_ ) + "* " * len(A_ ) ) return F"""SkipList(level={self.level})\n""" + "\n".join(A_ ) def __iter__( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.head while len(node.forward ) != 0: yield node.forward[0].key UpperCamelCase : Union[str, Any] = node.forward[0] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = 1 while random() < self.p and level < self.max_level: level += 1 return level def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[str] = [] UpperCamelCase : List[Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: UpperCamelCase : str = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(A_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : str = self._locate_node(A_ ) if node is not None: for i, update_node in enumerate(A_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: UpperCamelCase : Tuple = node.forward[i] else: UpperCamelCase : List[Any] = update_node.forward[:i] def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Optional[int] = self._locate_node(A_ ) if node is not None: UpperCamelCase : Union[str, Any] = value else: UpperCamelCase : Dict = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , A_ ): update_vector.append(self.head ) UpperCamelCase : Optional[int] = level UpperCamelCase : Dict = Node(A_ , A_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(A_ ) else: UpperCamelCase : List[Any] = new_node def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Union[str, Any] = self._locate_node(A_ ) if node is not None: return node.value return None def A_ ( ) -> List[Any]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) UpperCamelCase : Optional[int] = skip_list.head UpperCamelCase : List[str] = {} while node.level != 0: UpperCamelCase : str = node.forward[0] UpperCamelCase : Optional[int] = node.value assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def A_ ( ) -> List[Any]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) UpperCamelCase : Dict = skip_list.head UpperCamelCase : Tuple = {} while node.level != 0: UpperCamelCase : List[str] = node.forward[0] UpperCamelCase : Dict = node.value if len(_lowerCAmelCase ) != 4: print() assert len(_lowerCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() assert skip_list.find("Some key" ) is None def A_ ( ) -> Tuple: UpperCamelCase : Optional[int] = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def A_ ( ) -> Dict: UpperCamelCase : Optional[int] = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def A_ ( ) -> Dict: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[str]: UpperCamelCase : int = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def A_ ( ) -> List[Any]: UpperCamelCase : List[Any] = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 142 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_lowerCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_lowerCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def A_ ( ) -> Union[str, Any]: def is_sorted(_lowerCAmelCase ): return all(next_item >= item for item, next_item in zip(_lowerCAmelCase , lst[1:] ) ) UpperCamelCase : int = SkipList() for i in range(10 ): skip_list.insert(_lowerCAmelCase , _lowerCAmelCase ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_lowerCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_lowerCAmelCase ) ) def A_ ( ) -> Tuple: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def A_ ( ) -> List[str]: UpperCamelCase : Optional[int] = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
38
1
from math import pi def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> float: return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
38
from PIL import Image def A_ ( _lowerCAmelCase ) -> Image: UpperCamelCase , UpperCamelCase : List[Any] = image.size UpperCamelCase : Union[str, Any] = 0 UpperCamelCase : List[str] = image.load() for i in range(_lowerCAmelCase ): for j in range(_lowerCAmelCase ): UpperCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(_lowerCAmelCase ): for i in range(_lowerCAmelCase ): UpperCamelCase : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
38
1
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor __lowerCamelCase : int = logging.get_logger(__name__) class A__ ( __snake_case ): def __init__( self , *A_ , **A_ ): '''simple docstring''' warnings.warn( "The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use GLPNImageProcessor instead." , A_ , ) super().__init__(*A_ , **A_ )
38
from math import loga def A_ ( _lowerCAmelCase ) -> int: if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
1
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed __lowerCamelCase : Union[str, Any] = logging.getLogger(__name__) def A_ ( _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=16 , _lowerCAmelCase = 10 , _lowerCAmelCase = 2 ) -> Dict: def get_dataset(_lowerCAmelCase ): UpperCamelCase : str = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(_lowerCAmelCase , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) UpperCamelCase : Tuple = get_dataset(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = get_dataset(_lowerCAmelCase ) UpperCamelCase : int = DataLoader(_lowerCAmelCase , shuffle=_lowerCAmelCase , batch_size=_lowerCAmelCase , num_workers=4 ) UpperCamelCase : Union[str, Any] = DataLoader(_lowerCAmelCase , shuffle=_lowerCAmelCase , batch_size=_lowerCAmelCase , num_workers=4 ) return (train_dataloader, valid_dataloader) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ) -> Optional[Any]: UpperCamelCase : List[str] = [] for epoch in range(_lowerCAmelCase ): # Train quickly model.train() for batch in dataloader: UpperCamelCase , UpperCamelCase : str = batch UpperCamelCase : int = model(_lowerCAmelCase ) UpperCamelCase : Tuple = torch.nn.functional.mse_loss(_lowerCAmelCase , _lowerCAmelCase ) accelerator.backward(_lowerCAmelCase ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class A__ ( nn.Module ): def __init__( self ): '''simple docstring''' super().__init__() UpperCamelCase : int = nn.Parameter(torch.randn(1 ) ) UpperCamelCase : int = nn.Parameter(torch.randn(1 ) ) def __UpperCamelCase( self , A_ ): '''simple docstring''' return x * self.a + self.b class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase : List[Any] = DummyModel() UpperCamelCase : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase , UpperCamelCase : int = dummy_dataloaders() UpperCamelCase : Union[str, Any] = ProjectConfiguration(total_limit=1 , project_dir=A_ , automatic_checkpoint_naming=A_ ) # Train baseline UpperCamelCase : Union[str, Any] = Accelerator(project_config=A_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : Dict = accelerator.prepare( A_ , A_ , A_ , A_ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def __UpperCamelCase( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase : List[Any] = DummyModel() UpperCamelCase : Any = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase , UpperCamelCase : Dict = dummy_dataloaders() # Train baseline UpperCamelCase : List[str] = Accelerator() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = accelerator.prepare( A_ , A_ , A_ , A_ ) # Save initial UpperCamelCase : Tuple = os.path.join(A_ , "initial" ) accelerator.save_state(A_ ) ((UpperCamelCase) , (UpperCamelCase)) : Any = model.a.item(), model.b.item() UpperCamelCase : Optional[int] = optimizer.state_dict() UpperCamelCase : str = train(3 , A_ , A_ , A_ , A_ ) ((UpperCamelCase) , (UpperCamelCase)) : Tuple = model.a.item(), model.b.item() UpperCamelCase : str = optimizer.state_dict() # Train partially set_seed(42 ) UpperCamelCase : Optional[int] = DummyModel() UpperCamelCase : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase , UpperCamelCase : Any = dummy_dataloaders() UpperCamelCase : List[str] = Accelerator() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : Optional[Any] = accelerator.prepare( A_ , A_ , A_ , A_ ) accelerator.load_state(A_ ) ((UpperCamelCase) , (UpperCamelCase)) : Optional[Any] = model.a.item(), model.b.item() UpperCamelCase : List[str] = optimizer.state_dict() self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) UpperCamelCase : Tuple = train(2 , A_ , A_ , A_ , A_ ) # Save everything UpperCamelCase : Optional[int] = os.path.join(A_ , "checkpoint" ) accelerator.save_state(A_ ) # Load everything back in and make sure all states work accelerator.load_state(A_ ) test_rands += train(1 , A_ , A_ , A_ , A_ ) ((UpperCamelCase) , (UpperCamelCase)) : Any = model.a.item(), model.b.item() UpperCamelCase : Any = optimizer.state_dict() self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase : Dict = DummyModel() UpperCamelCase : str = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase , UpperCamelCase : int = dummy_dataloaders() UpperCamelCase : Dict = ProjectConfiguration(automatic_checkpoint_naming=A_ ) # Train baseline UpperCamelCase : int = Accelerator(project_dir=A_ , project_config=A_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : str = accelerator.prepare( A_ , A_ , A_ , A_ ) # Save initial accelerator.save_state() ((UpperCamelCase) , (UpperCamelCase)) : Optional[Any] = model.a.item(), model.b.item() UpperCamelCase : List[str] = optimizer.state_dict() UpperCamelCase : Tuple = train(3 , A_ , A_ , A_ , A_ ) ((UpperCamelCase) , (UpperCamelCase)) : List[str] = model.a.item(), model.b.item() UpperCamelCase : Any = optimizer.state_dict() # Train partially set_seed(42 ) UpperCamelCase : List[Any] = DummyModel() UpperCamelCase : Any = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase , UpperCamelCase : Tuple = dummy_dataloaders() UpperCamelCase : Dict = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=A_ ) UpperCamelCase : int = Accelerator(project_dir=A_ , project_config=A_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = accelerator.prepare( A_ , A_ , A_ , A_ ) accelerator.load_state(os.path.join(A_ , "checkpoints" , "checkpoint_0" ) ) ((UpperCamelCase) , (UpperCamelCase)) : Union[str, Any] = model.a.item(), model.b.item() UpperCamelCase : Optional[int] = optimizer.state_dict() self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) UpperCamelCase : str = train(2 , A_ , A_ , A_ , A_ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A_ , "checkpoints" , "checkpoint_1" ) ) test_rands += train(1 , A_ , A_ , A_ , A_ ) ((UpperCamelCase) , (UpperCamelCase)) : List[str] = model.a.item(), model.b.item() UpperCamelCase : Any = optimizer.state_dict() self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) self.assertEqual(A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = torch.tensor([1, 2, 3] ) UpperCamelCase : Any = torch.tensor([2, 3, 4] ) UpperCamelCase : Dict = DummyModel() UpperCamelCase : Union[str, Any] = torch.optim.Adam(net.parameters() ) UpperCamelCase : List[str] = Accelerator() with self.assertRaises(A_ ) as ve: accelerator.register_for_checkpointing(A_ , A_ , A_ , A_ ) UpperCamelCase : List[str] = str(ve.exception ) self.assertTrue("Item at index 0" in message ) self.assertTrue("Item at index 1" in message ) self.assertFalse("Item at index 2" in message ) self.assertFalse("Item at index 3" in message ) def __UpperCamelCase( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase : Any = DummyModel() UpperCamelCase : Optional[int] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) UpperCamelCase : Dict = torch.optim.lr_scheduler.StepLR(A_ , step_size=1 , gamma=0.99 ) UpperCamelCase , UpperCamelCase : Optional[int] = dummy_dataloaders() UpperCamelCase : Any = ProjectConfiguration(automatic_checkpoint_naming=A_ ) # Train baseline UpperCamelCase : Optional[Any] = Accelerator(project_dir=A_ , project_config=A_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : int = accelerator.prepare( A_ , A_ , A_ , A_ , A_ ) # Save initial accelerator.save_state() UpperCamelCase : Union[str, Any] = scheduler.state_dict() train(3 , A_ , A_ , A_ , A_ , A_ ) self.assertNotEqual(A_ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A_ , "checkpoints" , "checkpoint_0" ) ) self.assertEqual(A_ , scheduler.state_dict() ) def __UpperCamelCase( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase : List[str] = DummyModel() UpperCamelCase : int = ProjectConfiguration(automatic_checkpoint_naming=A_ , total_limit=2 ) # Train baseline UpperCamelCase : Optional[int] = Accelerator(project_dir=A_ , project_config=A_ ) UpperCamelCase : Tuple = accelerator.prepare(A_ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(A_ , "checkpoints" , "checkpoint_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(A_ , "checkpoints" , "checkpoint_9" ) ) ) self.assertTrue(os.path.exists(os.path.join(A_ , "checkpoints" , "checkpoint_10" ) ) ) @require_cuda def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = ["torchrun", F"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(A_ , env=os.environ.copy() ) if __name__ == "__main__": __lowerCamelCase : str = """/tmp/accelerate/state_checkpointing""" __lowerCamelCase : int = DummyModel() __lowerCamelCase : Optional[Any] = torch.optim.Adam(params=model.parameters(), lr=1E-3) __lowerCamelCase : Any = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.9_9) __lowerCamelCase , __lowerCamelCase : Dict = dummy_dataloaders() __lowerCamelCase : Any = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline __lowerCamelCase : Any = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision="""no""") if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Any = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) __lowerCamelCase , __lowerCamelCase : List[Any] = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: __lowerCamelCase : int = group["""params"""][0].device break assert param_device.type == accelerator.device.type __lowerCamelCase : List[str] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""cpu""") for group in optimizer.param_groups: __lowerCamelCase : Optional[Any] = group["""params"""][0].device break assert ( param_device.type == torch.device("""cpu""").type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""on_device""") for group in optimizer.param_groups: __lowerCamelCase : Optional[int] = group["""params"""][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match="""Unsupported optimizer map location passed"""): accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""invalid""") accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
38
from __future__ import annotations __lowerCamelCase : Optional[int] = """Muhammad Umer Farooq""" __lowerCamelCase : Tuple = """MIT""" __lowerCamelCase : Optional[int] = """1.0.0""" __lowerCamelCase : int = """Muhammad Umer Farooq""" __lowerCamelCase : Optional[int] = """contact@muhammadumerfarooq.me""" __lowerCamelCase : Dict = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __snake_case ): def __init__( self , A_ ): '''simple docstring''' super().__init__() UpperCamelCase : list[str] = [] UpperCamelCase : str = domain def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCamelCase : Any = parse.urljoin(self.domain , A_ ) self.urls.append(A_ ) def A_ ( _lowerCAmelCase ) -> str: return ".".join(get_sub_domain_name(_lowerCAmelCase ).split("." )[-2:] ) def A_ ( _lowerCAmelCase ) -> str: return parse.urlparse(_lowerCAmelCase ).netloc def A_ ( _lowerCAmelCase = "https://github.com" ) -> list[str]: UpperCamelCase : int = get_domain_name(_lowerCAmelCase ) # Initialize the parser UpperCamelCase : str = Parser(_lowerCAmelCase ) try: # Open URL UpperCamelCase : int = requests.get(_lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCamelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCamelCase : Optional[Any] = requests.get(_lowerCAmelCase ) # Get the valid email. UpperCamelCase : Optional[int] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = emails_from_url("""https://github.com""") print(f"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
38
1
from __future__ import annotations from typing import Any class A__ : def __init__( self , A_ ): '''simple docstring''' UpperCamelCase : str = num_of_nodes UpperCamelCase : list[list[int]] = [] UpperCamelCase : dict[int, int] = {} def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' self.m_edges.append([u_node, v_node, weight] ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def __UpperCamelCase( self , A_ ): '''simple docstring''' if self.m_component[u_node] != u_node: for k in self.m_component: UpperCamelCase : int = self.find_component(A_ ) def __UpperCamelCase( self , A_ , A_ , A_ ): '''simple docstring''' if component_size[u_node] <= component_size[v_node]: UpperCamelCase : Any = v_node component_size[v_node] += component_size[u_node] self.set_component(A_ ) elif component_size[u_node] >= component_size[v_node]: UpperCamelCase : List[Any] = self.find_component(A_ ) component_size[u_node] += component_size[v_node] self.set_component(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = [] UpperCamelCase : Tuple = 0 UpperCamelCase : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) UpperCamelCase : List[str] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = edge UpperCamelCase : List[str] = self.m_component[u] UpperCamelCase : Optional[Any] = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): UpperCamelCase : List[str] = [u, v, w] for edge in minimum_weight_edge: if isinstance(A_ , A_ ): UpperCamelCase , UpperCamelCase , UpperCamelCase : Optional[int] = edge UpperCamelCase : Dict = self.m_component[u] UpperCamelCase : Optional[Any] = self.m_component[v] if u_component != v_component: mst_weight += w self.union(A_ , A_ , A_ ) print(F"""Added edge [{u} - {v}]\nAdded weight: {w}\n""" ) num_of_components -= 1 UpperCamelCase : Optional[Any] = [-1] * self.m_num_of_nodes print(F"""The total weight of the minimal spanning tree is: {mst_weight}""" ) def A_ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
38
from __future__ import annotations def A_ ( _lowerCAmelCase ) -> list[int]: UpperCamelCase : Optional[Any] = [True] * limit UpperCamelCase : Optional[Any] = False UpperCamelCase : List[str] = False UpperCamelCase : Tuple = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCamelCase : Optional[Any] = i * 2 while index < limit: UpperCamelCase : int = False UpperCamelCase : Optional[int] = index + i UpperCamelCase : Any = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def A_ ( _lowerCAmelCase = 100_0000 ) -> int: UpperCamelCase : Union[str, Any] = prime_sieve(_lowerCAmelCase ) UpperCamelCase : List[str] = 0 UpperCamelCase : Union[str, Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCamelCase : Dict = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCamelCase : int = j - i UpperCamelCase : Dict = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
38
1
__lowerCamelCase : Optional[int] = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] __lowerCamelCase : Dict = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] __lowerCamelCase : str = { 0: """Sunday""", 1: """Monday""", 2: """Tuesday""", 3: """Wednesday""", 4: """Thursday""", 5: """Friday""", 6: """Saturday""", } def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> str: assert len(str(_lowerCAmelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: UpperCamelCase : int = year // 100 UpperCamelCase : Optional[Any] = (5 * (century % 4) + 2) % 7 UpperCamelCase : str = year % 100 UpperCamelCase : List[str] = centurian % 12 UpperCamelCase : Union[str, Any] = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 UpperCamelCase : List[str] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) UpperCamelCase : Dict = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
38
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __snake_case ): def __init__( self , A_ , A_ = None , A_ = None , A_ = False , A_ = False , A_ = None , A_ = None , **A_ , ): '''simple docstring''' super().__init__( features=A_ , cache_dir=A_ , keep_in_memory=A_ , streaming=A_ , num_proc=A_ , **A_ , ) UpperCamelCase : Optional[int] = Generator( cache_dir=A_ , features=A_ , generator=A_ , gen_kwargs=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' if self.streaming: UpperCamelCase : Optional[Any] = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: UpperCamelCase : Union[str, Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , num_proc=self.num_proc , ) UpperCamelCase : int = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset
38
1
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : Optional[Any] = """.""" # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : int = [ """Assert""", """AssignVariableOp""", """EmptyTensorList""", """MergeV2Checkpoints""", """ReadVariableOp""", """ResourceGather""", """RestoreV2""", """SaveV2""", """ShardedFilename""", """StatefulPartitionedCall""", """StaticRegexFullMatch""", """VarHandleOp""", ] def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[str]: UpperCamelCase : int = SavedModel() UpperCamelCase : Any = [] with open(os.path.join(_lowerCAmelCase , "utils" , "tf_ops" , "onnx.json" ) ) as f: UpperCamelCase : Optional[Any] = json.load(_lowerCAmelCase )["opsets"] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(_lowerCAmelCase )] ) with open(_lowerCAmelCase , "rb" ) as f: saved_model.ParseFromString(f.read() ) UpperCamelCase : int = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want UpperCamelCase : Union[str, Any] = sorted(_lowerCAmelCase ) UpperCamelCase : Any = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(_lowerCAmelCase ) if strict and len(_lowerCAmelCase ) > 0: raise Exception(F"""Found the following incompatible ops for the opset {opset}:\n""" + incompatible_ops ) elif len(_lowerCAmelCase ) > 0: print(F"""Found the following incompatible ops for the opset {opset}:""" ) print(*_lowerCAmelCase , sep="\n" ) else: print(F"""The saved model {saved_model_path} can properly be converted with ONNX.""" ) if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("""--saved_model_path""", help="""Path of the saved model to check (the .pb file).""") parser.add_argument( """--opset""", default=12, type=int, help="""The ONNX opset against which the model has to be tested.""" ) parser.add_argument( """--framework""", choices=["""onnx"""], default="""onnx""", help="""Frameworks against which to test the saved model.""" ) parser.add_argument( """--strict""", action="""store_true""", help="""Whether make the checking strict (raise errors) or not (raise warnings)""" ) __lowerCamelCase : str = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
38
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # picklable for multiprocessing return x.sum() def A_ ( _lowerCAmelCase ) -> Optional[Any]: # picklable for multiprocessing return i + 1 @dataclass class A__ : _UpperCAmelCase :int _UpperCAmelCase :str class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = {} UpperCamelCase : Optional[Any] = [] UpperCamelCase : List[Any] = 1 UpperCamelCase : Tuple = [1, 2] UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : Optional[Any] = {"a": [1, 2], "b": [3, 4]} UpperCamelCase : Any = {"a": {"1": 1}, "b": 2} UpperCamelCase : List[str] = {"a": 1, "b": 2, "c": 3, "d": 4} UpperCamelCase : Dict = {} UpperCamelCase : Any = [] UpperCamelCase : Any = 2 UpperCamelCase : Any = [2, 3] UpperCamelCase : Optional[Any] = {"a": 2, "b": 3} UpperCamelCase : List[Any] = {"a": [2, 3], "b": [4, 5]} UpperCamelCase : Tuple = {"a": {"1": 2}, "b": 3} UpperCamelCase : Dict = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ ) , A_ ) UpperCamelCase : List[str] = 2 self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) self.assertEqual(map_nested(A_ , A_ , num_proc=A_ ) , A_ ) UpperCamelCase : List[str] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} UpperCamelCase : int = {"a": 2, "b": 0, "c": 2} UpperCamelCase : Union[str, Any] = { "a": np.eye(2 ).astype(A_ ), "b": np.zeros(3 ).astype(A_ ), "c": np.ones(2 ).astype(A_ ), } self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ) , A_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(A_ , A_ , map_numpy=A_ , num_proc=A_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(A_ ): # can't pickle a local lambda map_nested(lambda A_ : x + 1 , A_ , num_proc=A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = {"a": 1, "b": 2} UpperCamelCase : List[Any] = {"a": 3, "b": 4} UpperCamelCase : Tuple = {"a": 5, "b": 6} UpperCamelCase : Union[str, Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(A_ , A_ , A_ ) ) , A_ ) def __UpperCamelCase( self ): '''simple docstring''' class A__ : _UpperCAmelCase :str = 'bar' UpperCamelCase : List[Any] = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(A_ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: UpperCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(_lowerCAmelCase )} UpperCamelCase : List[str] = map_nested(lambda _lowerCAmelCase : x + 10 , _lowerCAmelCase , num_proc=_lowerCAmelCase , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A__ ( __snake_case ): @require_tf def __UpperCamelCase( self ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers UpperCamelCase : int = layers.Dense(2 ) def gen_random_output(): UpperCamelCase : Optional[Any] = tf.random.uniform((1, 3) ) return model(A_ ).numpy() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : List[Any] = gen_random_output() with temp_seed(42 , set_tensorflow=A_ ): UpperCamelCase : Dict = gen_random_output() UpperCamelCase : Optional[int] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def __UpperCamelCase( self ): '''simple docstring''' import torch def gen_random_output(): UpperCamelCase : Optional[Any] = torch.nn.Linear(3 , 2 ) UpperCamelCase : Dict = torch.rand(1 , 3 ) return model(A_ ).detach().numpy() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Dict = gen_random_output() with temp_seed(42 , set_pytorch=A_ ): UpperCamelCase : Optional[int] = gen_random_output() UpperCamelCase : List[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def __UpperCamelCase( self ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() with temp_seed(42 ): UpperCamelCase : Optional[Any] = gen_random_output() UpperCamelCase : Optional[Any] = gen_random_output() np.testing.assert_equal(A_ , A_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" , [{}] ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Optional[Any] = NestedDataStructure(_lowerCAmelCase ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" , [ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] , ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = NestedDataStructure(_lowerCAmelCase ).flatten() assert output == expected_output def A_ ( ) -> List[Any]: UpperCamelCase : str = A(x=1 , y="foobar" ) UpperCamelCase : Tuple = {"x": 1, "y": "foobar"} assert asdict(_lowerCAmelCase ) == expected_output UpperCamelCase : List[str] = {"a": {"b": A(x=10 , y="foo" )}, "c": [A(x=20 , y="bar" )]} UpperCamelCase : Tuple = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_lowerCAmelCase ) == expected_output with pytest.raises(_lowerCAmelCase ): asdict([1, A(x=10 , y="foo" )] ) def A_ ( _lowerCAmelCase ) -> Tuple: return text.split() def A_ ( _lowerCAmelCase ) -> Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def A_ ( ) -> str: with Pool(2 ) as pool: UpperCamelCase : List[str] = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: UpperCamelCase : Dict = list(iflatmap_unordered(_lowerCAmelCase , _split_text , kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(_lowerCAmelCase ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: UpperCamelCase : Any = [] for yield_time, content in iflatmap_unordered( _lowerCAmelCase , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_lowerCAmelCase ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(_lowerCAmelCase ) == 4
38
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=A_ , ) assert hasattr(self , "env" ) def __UpperCamelCase( self , A_=1 ): '''simple docstring''' return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=A_ , instance_type=self.instance_type , debugger_hook_config=A_ , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def __UpperCamelCase( self , A_ ): '''simple docstring''' TrainingJobAnalytics(A_ ).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.create_estimator() # run training estimator.fit() # result dataframe UpperCamelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCamelCase : List[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCamelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCamelCase : Optional[Any] = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , A_ )
38
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): _UpperCAmelCase :Tuple = ['note_seq'] def __init__( self , *A_ , **A_ ): '''simple docstring''' requires_backends(self , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] ) @classmethod def __UpperCamelCase( cls , *A_ , **A_ ): '''simple docstring''' requires_backends(cls , ["note_seq"] )
38
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() __lowerCamelCase : Optional[Any] = logging.get_logger(__name__) __lowerCamelCase : str = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn.grep_linear""": """encoder.layers.*.attention.gru_rel_pos_linear""", """self_attn.relative_attention_bias""": """encoder.layers.*.attention.rel_attn_embed""", """self_attn.grep_a""": """encoder.layers.*.attention.gru_rel_pos_const""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } __lowerCamelCase : str = [ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Dict: for attribute in key.split("." ): UpperCamelCase : Tuple = getattr(_lowerCAmelCase , _lowerCAmelCase ) if weight_type is not None: UpperCamelCase : Dict = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape else: UpperCamelCase : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCamelCase : Optional[Any] = value elif weight_type == "weight_g": UpperCamelCase : Dict = value elif weight_type == "weight_v": UpperCamelCase : Union[str, Any] = value elif weight_type == "bias": UpperCamelCase : Optional[Any] = value else: UpperCamelCase : List[str] = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : Optional[int] = [] UpperCamelCase : Tuple = fairseq_model.state_dict() UpperCamelCase : Any = hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCamelCase : Optional[int] = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == "group" , ) UpperCamelCase : Dict = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: UpperCamelCase : Any = True if "*" in mapped_key: UpperCamelCase : Any = name.split(_lowerCAmelCase )[0].split("." )[-2] UpperCamelCase : Optional[Any] = mapped_key.replace("*" , _lowerCAmelCase ) if "weight_g" in name: UpperCamelCase : Optional[Any] = "weight_g" elif "weight_v" in name: UpperCamelCase : Any = "weight_v" elif "bias" in name and "relative_attention_bias" not in name: UpperCamelCase : Optional[int] = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCamelCase : Tuple = "weight" else: UpperCamelCase : Dict = None set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) continue if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(F"""Unused weights: {unused_weights}""" ) def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: UpperCamelCase : str = full_name.split("conv_layers." )[-1] UpperCamelCase : Optional[int] = name.split("." ) UpperCamelCase : str = int(items[0] ) UpperCamelCase : Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCamelCase : str = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCamelCase : int = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) UpperCamelCase : str = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCamelCase : Optional[Any] = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ) -> List[str]: # load the pre-trained checkpoints UpperCamelCase : Optional[Any] = torch.load(_lowerCAmelCase ) UpperCamelCase : Optional[Any] = WavLMConfigOrig(checkpoint["cfg"] ) UpperCamelCase : Dict = WavLMOrig(_lowerCAmelCase ) model.load_state_dict(checkpoint["model"] ) model.eval() if config_path is not None: UpperCamelCase : Optional[int] = WavLMConfig.from_pretrained(_lowerCAmelCase ) else: UpperCamelCase : Tuple = WavLMConfig() UpperCamelCase : List[Any] = WavLMModel(_lowerCAmelCase ) recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase ) hf_wavlm.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Dict = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") __lowerCamelCase : Tuple = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
38
import math import tensorflow as tf from packaging import version def A_ ( _lowerCAmelCase ) -> Any: UpperCamelCase : List[Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def A_ ( _lowerCAmelCase ) -> Dict: UpperCamelCase : Union[str, Any] = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(math.pi , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : int = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : str = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def A_ ( _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Tuple = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(0.044_715 , x.dtype ) UpperCamelCase : Optional[Any] = tf.cast(0.7_978_845_608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def A_ ( _lowerCAmelCase ) -> Optional[Any]: UpperCamelCase : Any = tf.convert_to_tensor(_lowerCAmelCase ) UpperCamelCase : List[Any] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def A_ ( _lowerCAmelCase ) -> List[Any]: return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def A_ ( _lowerCAmelCase , _lowerCAmelCase=-1 ) -> str: UpperCamelCase , UpperCamelCase : List[Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def A_ ( _lowerCAmelCase ) -> Any: return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) __lowerCamelCase : Optional[int] = tf.keras.activations.gelu __lowerCamelCase : int = approximate_gelu_wrap else: __lowerCamelCase : List[Any] = _gelu __lowerCamelCase : Optional[Any] = _gelu_new __lowerCamelCase : Any = { """gelu""": gelu, """gelu_10""": gelu_aa, """gelu_fast""": gelu_fast, """gelu_new""": gelu_new, """glu""": glu, """mish""": mish, """quick_gelu""": quick_gelu, """relu""": tf.keras.activations.relu, """sigmoid""": tf.keras.activations.sigmoid, """silu""": tf.keras.activations.swish, """swish""": tf.keras.activations.swish, """tanh""": tf.keras.activations.tanh, } def A_ ( _lowerCAmelCase ) -> Optional[Any]: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
38
1
import numpy as np __lowerCamelCase : Union[str, Any] = [ ["""a""", """b""", """c""", """d""", """e"""], ["""f""", """g""", """h""", """i""", """k"""], ["""l""", """m""", """n""", """o""", """p"""], ["""q""", """r""", """s""", """t""", """u"""], ["""v""", """w""", """x""", """y""", """z"""], ] class A__ : def __init__( self ): '''simple docstring''' UpperCamelCase : Dict = np.array(A_ ) def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Any = np.where(letter == self.SQUARE ) UpperCamelCase : Union[str, Any] = np.concatenate([indexa + 1, indexa + 1] ) return indexes def __UpperCamelCase( self , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.SQUARE[indexa - 1, indexa - 1] return letter def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = message.lower() UpperCamelCase : int = message.replace(" " , "" ) UpperCamelCase : Optional[int] = message.replace("j" , "i" ) UpperCamelCase : int = np.empty((2, len(A_ )) ) for letter_index in range(len(A_ ) ): UpperCamelCase : Optional[int] = self.letter_to_numbers(message[letter_index] ) UpperCamelCase : Optional[int] = numbers[0] UpperCamelCase : List[Any] = numbers[1] UpperCamelCase : Any = first_step.reshape(2 * len(A_ ) ) UpperCamelCase : List[Any] = "" for numbers_index in range(len(A_ ) ): UpperCamelCase : int = int(second_step[numbers_index * 2] ) UpperCamelCase : int = int(second_step[(numbers_index * 2) + 1] ) UpperCamelCase : List[Any] = self.numbers_to_letter(A_ , A_ ) UpperCamelCase : str = encoded_message + letter return encoded_message def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = message.lower() message.replace(" " , "" ) UpperCamelCase : Any = np.empty(2 * len(A_ ) ) for letter_index in range(len(A_ ) ): UpperCamelCase : Tuple = self.letter_to_numbers(message[letter_index] ) UpperCamelCase : Optional[int] = numbers[0] UpperCamelCase : List[str] = numbers[1] UpperCamelCase : List[str] = first_step.reshape((2, len(A_ )) ) UpperCamelCase : int = "" for numbers_index in range(len(A_ ) ): UpperCamelCase : str = int(second_step[0, numbers_index] ) UpperCamelCase : Dict = int(second_step[1, numbers_index] ) UpperCamelCase : str = self.numbers_to_letter(A_ , A_ ) UpperCamelCase : Optional[int] = decoded_message + letter return decoded_message
38
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :str = KandinskyVaaPipeline _UpperCAmelCase :str = [ 'image_embeds', 'negative_image_embeds', ] _UpperCAmelCase :str = ['image_embeds', 'negative_image_embeds'] _UpperCAmelCase :List[str] = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _UpperCAmelCase :List[str] = False @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return 32 @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim @property def __UpperCamelCase( self ): '''simple docstring''' return self.time_input_dim * 4 @property def __UpperCamelCase( self ): '''simple docstring''' return 100 @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : List[str] = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } UpperCamelCase : Dict = UNetaDConditionModel(**A_ ) return model @property def __UpperCamelCase( self ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCamelCase( self ): '''simple docstring''' torch.manual_seed(0 ) UpperCamelCase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.dummy_unet UpperCamelCase : Optional[Any] = self.dummy_movq UpperCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=A_ , set_alpha_to_one=A_ , steps_offset=1 , prediction_type="epsilon" , thresholding=A_ , ) UpperCamelCase : Tuple = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def __UpperCamelCase( self , A_ , A_=0 ): '''simple docstring''' UpperCamelCase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A_ ) ).to(A_ ) UpperCamelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A_ ) if str(A_ ).startswith("mps" ): UpperCamelCase : Optional[Any] = torch.manual_seed(A_ ) else: UpperCamelCase : List[Any] = torch.Generator(device=A_ ).manual_seed(A_ ) UpperCamelCase : Optional[int] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = "cpu" UpperCamelCase : List[str] = self.get_dummy_components() UpperCamelCase : Tuple = self.pipeline_class(**A_ ) UpperCamelCase : List[str] = pipe.to(A_ ) pipe.set_progress_bar_config(disable=A_ ) UpperCamelCase : Dict = pipe(**self.get_dummy_inputs(A_ ) ) UpperCamelCase : Optional[int] = output.images UpperCamelCase : int = pipe( **self.get_dummy_inputs(A_ ) , return_dict=A_ , )[0] UpperCamelCase : Tuple = image[0, -3:, -3:, -1] UpperCamelCase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase : int = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class A__ ( unittest.TestCase ): def __UpperCamelCase( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) UpperCamelCase : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(A_ ) UpperCamelCase : Dict = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) UpperCamelCase : Tuple = pipeline.to(A_ ) pipeline.set_progress_bar_config(disable=A_ ) UpperCamelCase : str = "red cat, 4k photo" UpperCamelCase : str = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase , UpperCamelCase : Tuple = pipe_prior( A_ , generator=A_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() UpperCamelCase : int = torch.Generator(device="cuda" ).manual_seed(0 ) UpperCamelCase : Tuple = pipeline( image_embeds=A_ , negative_image_embeds=A_ , generator=A_ , num_inference_steps=100 , output_type="np" , ) UpperCamelCase : Union[str, Any] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A_ , A_ )
38
1
from math import factorial __lowerCamelCase : List[Any] = {str(d): factorial(d) for d in range(10)} def A_ ( _lowerCAmelCase ) -> int: return sum(DIGIT_FACTORIAL[d] for d in str(_lowerCAmelCase ) ) def A_ ( ) -> int: UpperCamelCase : Optional[Any] = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , _lowerCAmelCase ) if sum_of_digit_factorial(_lowerCAmelCase ) == i ) if __name__ == "__main__": print(f"""{solution() = }""")
38
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def A_ ( ) -> Dict: UpperCamelCase : Tuple = ArgumentParser( description=( "PyTorch TPU distributed training launch " "helper utility that will spawn up " "multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=_lowerCAmelCase , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=_lowerCAmelCase , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=_lowerCAmelCase ) return parser.parse_args() def A_ ( ) -> Optional[int]: UpperCamelCase : Tuple = parse_args() # Import training_script as a module. UpperCamelCase : Union[str, Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase : List[Any] = script_fpath.stem UpperCamelCase : Optional[Any] = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv UpperCamelCase : List[Any] = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
38
1
import os def A_ ( ) -> List[str]: with open(os.path.dirname(_lowerCAmelCase ) + "/grid.txt" ) as f: UpperCamelCase : int = [] # noqa: E741 for _ in range(20 ): l.append([int(_lowerCAmelCase ) for x in f.readline().split()] ) UpperCamelCase : str = 0 # right for i in range(20 ): for j in range(17 ): UpperCamelCase : List[Any] = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: UpperCamelCase : Any = temp # down for i in range(17 ): for j in range(20 ): UpperCamelCase : List[str] = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: UpperCamelCase : List[Any] = temp # diagonal 1 for i in range(17 ): for j in range(17 ): UpperCamelCase : Union[str, Any] = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: UpperCamelCase : Optional[int] = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): UpperCamelCase : Dict = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: UpperCamelCase : int = temp return maximum if __name__ == "__main__": print(solution())
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCamelCase : Union[str, Any] = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __lowerCamelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : str = 13 UpperCamelCase : int = 7 UpperCamelCase : str = True UpperCamelCase : Dict = True UpperCamelCase : str = True UpperCamelCase : Tuple = True UpperCamelCase : List[str] = 99 UpperCamelCase : Optional[Any] = 384 UpperCamelCase : Tuple = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Dict = 37 UpperCamelCase : Any = "gelu" UpperCamelCase : List[Any] = 0.1 UpperCamelCase : int = 0.1 UpperCamelCase : Tuple = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : int = 2 UpperCamelCase : Dict = 0.02 UpperCamelCase : Optional[Any] = 3 UpperCamelCase : List[Any] = 4 UpperCamelCase : Dict = 128 UpperCamelCase : Optional[Any] = 2 UpperCamelCase : Optional[int] = 9 UpperCamelCase : Optional[int] = 1 UpperCamelCase : Union[str, Any] = None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : str = None if self.use_input_mask: UpperCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None if self.use_token_type_ids: UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : Any = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel(config=A_ ) UpperCamelCase : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : Any = model(A_ ) UpperCamelCase : int = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = TFConvBertForMaskedLM(config=A_ ) UpperCamelCase : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = TFConvBertForSequenceClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_choices UpperCamelCase : str = TFConvBertForMultipleChoice(config=A_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Dict = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Any = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : List[str] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCamelCase : Optional[Any] = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : str = TFConvBertForTokenClassification(config=A_ ) UpperCamelCase : List[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : str = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = TFConvBertForQuestionAnswering(config=A_ ) UpperCamelCase : Union[str, Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCamelCase : Union[str, Any] = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase :Any = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = TFConvBertModelTester(self ) UpperCamelCase : Dict = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Optional[Any] = True UpperCamelCase : Any = True if hasattr(A_ , "use_cache" ): UpperCamelCase : List[str] = True UpperCamelCase : List[Any] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Any = getattr(self.model_tester , "key_length" , A_ ) for model_class in self.all_model_classes: UpperCamelCase : List[Any] = self._prepare_for_class(A_ , A_ ) UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Optional[int] = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) UpperCamelCase : Union[str, Any] = os.path.join(A_ , "saved_model" , "1" ) UpperCamelCase : Dict = tf.keras.models.load_model(A_ ) UpperCamelCase : str = model(A_ ) if self.is_encoder_decoder: UpperCamelCase : Union[str, Any] = outputs["encoder_hidden_states"] UpperCamelCase : Any = outputs["encoder_attentions"] else: UpperCamelCase : Any = outputs["hidden_states"] UpperCamelCase : List[str] = outputs["attentions"] self.assertEqual(len(A_ ) , A_ ) UpperCamelCase : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase : Dict = True UpperCamelCase : int = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCamelCase : Optional[int] = getattr(self.model_tester , "key_length" , A_ ) UpperCamelCase : Optional[Any] = getattr(self.model_tester , "key_length" , A_ ) def check_decoder_attentions_output(A_ ): UpperCamelCase : Optional[Any] = len(A_ ) self.assertEqual(out_len % 2 , 0 ) UpperCamelCase : Any = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): UpperCamelCase : Dict = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True UpperCamelCase : List[Any] = False UpperCamelCase : Dict = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : List[str] = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Tuple = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase : Tuple = True UpperCamelCase : int = model_class(A_ ) UpperCamelCase : Dict = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine UpperCamelCase : Optional[int] = True UpperCamelCase : List[str] = True UpperCamelCase : Optional[int] = model_class(A_ ) UpperCamelCase : Optional[Any] = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCamelCase : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : List[str] = model(A_ )[0] UpperCamelCase : int = [1, 6, 768] self.assertEqual(output.shape , A_ ) UpperCamelCase : List[str] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
38
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class A__ ( unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=3 , A_=10 , A_=18 , A_=30 , A_=400 , A_=True , A_=None , A_=True , A_=[0.5, 0.5, 0.5] , A_=[0.5, 0.5, 0.5] , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[int] = size if size is not None else {"shortest_edge": 18} UpperCamelCase : Tuple = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCamelCase : Optional[Any] = parent UpperCamelCase : Optional[int] = batch_size UpperCamelCase : List[Any] = num_channels UpperCamelCase : Union[str, Any] = num_frames UpperCamelCase : Any = image_size UpperCamelCase : Tuple = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Any = do_resize UpperCamelCase : Tuple = size UpperCamelCase : List[Any] = do_normalize UpperCamelCase : Optional[int] = image_mean UpperCamelCase : Any = image_std UpperCamelCase : str = crop_size def __UpperCamelCase( self ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A__ ( __snake_case , unittest.TestCase ): _UpperCAmelCase :List[str] = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = VivitImageProcessingTester(self ) @property def __UpperCamelCase( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "do_center_crop" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCamelCase : Any = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : str = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase : str = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : Any = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for video in video_inputs: self.assertIsInstance(A_ , A_ ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCamelCase : Tuple = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCamelCase : List[Any] = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
38
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Any = logging.get_logger(__name__) __lowerCamelCase : int = { """google/mobilenet_v1_1.0_224""": """https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json""", """google/mobilenet_v1_0.75_192""": """https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json""", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class A__ ( __snake_case ): _UpperCAmelCase :List[Any] = 'mobilenet_v1' def __init__( self , A_=3 , A_=224 , A_=1.0 , A_=8 , A_="relu6" , A_=True , A_=0.9_99 , A_=0.02 , A_=0.0_01 , **A_ , ): '''simple docstring''' super().__init__(**A_ ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) UpperCamelCase : List[str] = num_channels UpperCamelCase : Any = image_size UpperCamelCase : Optional[Any] = depth_multiplier UpperCamelCase : Union[str, Any] = min_depth UpperCamelCase : int = hidden_act UpperCamelCase : Union[str, Any] = tf_padding UpperCamelCase : Optional[Any] = classifier_dropout_prob UpperCamelCase : int = initializer_range UpperCamelCase : Tuple = layer_norm_eps class A__ ( __snake_case ): _UpperCAmelCase :Any = version.parse('1.11' ) @property def __UpperCamelCase( self ): '''simple docstring''' return OrderedDict([("pixel_values", {0: "batch"})] ) @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def __UpperCamelCase( self ): '''simple docstring''' return 1e-4
38
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __lowerCamelCase : Dict = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } __lowerCamelCase : Tuple = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = VOCAB_FILES_NAMES _UpperCAmelCase :Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Optional[Any] = BlenderbotSmallTokenizer def __init__( self , A_=None , A_=None , A_="<|endoftext|>" , A_="<|endoftext|>" , A_="<|endoftext|>" , A_=False , A_=True , **A_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=A_ , merges=A_ , add_prefix_space=A_ , trim_offsets=A_ , ) , bos_token=A_ , eos_token=A_ , unk_token=A_ , **A_ , ) UpperCamelCase : Union[str, Any] = add_prefix_space def __UpperCamelCase( self , A_ , A_=None ): '''simple docstring''' UpperCamelCase : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase( self , A_ , A_ = None ): '''simple docstring''' UpperCamelCase : Tuple = [self.sep_token_id] UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
38
1
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class A__ ( __snake_case ): def __init__( self , A_ , A_ , A_ = None , A_ = None , A_ = False , **A_ , ): '''simple docstring''' super().__init__(features=A_ , cache_dir=A_ , keep_in_memory=A_ , **A_ ) UpperCamelCase : List[Any] = Sql( cache_dir=A_ , features=A_ , sql=A_ , con=A_ , **A_ , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = None UpperCamelCase : List[Any] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : Tuple = None self.builder.download_and_prepare( download_config=A_ , download_mode=A_ , verification_mode=A_ , base_path=A_ , ) # Build dataset for splits UpperCamelCase : Dict = self.builder.as_dataset( split="train" , verification_mode=A_ , in_memory=self.keep_in_memory ) return dataset class A__ : def __init__( self , A_ , A_ , A_ , A_ = None , A_ = None , **A_ , ): '''simple docstring''' if num_proc is not None and num_proc <= 0: raise ValueError(F"""num_proc {num_proc} must be an integer > 0.""" ) UpperCamelCase : Dict = dataset UpperCamelCase : List[Any] = name UpperCamelCase : Tuple = con UpperCamelCase : Any = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE UpperCamelCase : Dict = num_proc UpperCamelCase : str = to_sql_kwargs def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.to_sql_kwargs.pop("sql" , A_ ) UpperCamelCase : int = self.to_sql_kwargs.pop("con" , A_ ) UpperCamelCase : Union[str, Any] = self.to_sql_kwargs.pop("index" , A_ ) UpperCamelCase : List[Any] = self._write(index=A_ , **self.to_sql_kwargs ) return written def __UpperCamelCase( self , A_ ): '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = args UpperCamelCase : List[str] = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs UpperCamelCase : Tuple = query_table( table=self.dataset.data , key=slice(A_ , offset + self.batch_size ) , indices=self.dataset._indices , ) UpperCamelCase : Union[str, Any] = batch.to_pandas() UpperCamelCase : str = df.to_sql(self.name , self.con , index=A_ , **A_ ) return num_rows or len(A_ ) def __UpperCamelCase( self , A_ , **A_ ): '''simple docstring''' UpperCamelCase : Optional[Any] = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: UpperCamelCase , UpperCamelCase : List[str] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql , [(offset, index, to_sql_kwargs) for offset in range(0 , A_ , A_ )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += num_rows return written
38
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : int = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : str = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
38
1