code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): @property def A_ ( self : str ) -> Optional[Any]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def A_ ( self : Optional[Any] ) -> int: lowerCamelCase__ : Tuple = ort.SessionOptions() lowerCamelCase__ : int = False return options def A_ ( self : Optional[Any] ) -> List[str]: lowerCamelCase__ : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) lowerCamelCase__ : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) lowerCamelCase__ : List[str] = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , safety_checker=UpperCAmelCase , feature_extractor=UpperCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowerCamelCase__ : List[Any] = 'A red cat sitting on a park bench' lowerCamelCase__ : Dict = np.random.RandomState(0 ) lowerCamelCase__ : Optional[int] = pipe( prompt=UpperCAmelCase , image=UpperCAmelCase , mask_image=UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase , output_type='np' , ) lowerCamelCase__ : Any = output.images lowerCamelCase__ : Union[str, Any] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) lowerCamelCase__ : Union[str, Any] = np.array([0.2_5_1_4, 0.3_0_0_7, 0.3_5_1_7, 0.1_7_9_0, 0.2_3_8_2, 0.3_1_6_7, 0.1_9_4_4, 0.2_2_7_3, 0.2_4_6_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) lowerCamelCase__ : Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) lowerCamelCase__ : Tuple = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-inpainting' , subfolder='scheduler' , revision='onnx' ) lowerCamelCase__ : str = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , scheduler=UpperCAmelCase , safety_checker=UpperCAmelCase , feature_extractor=UpperCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowerCamelCase__ : str = 'A red cat sitting on a park bench' lowerCamelCase__ : Tuple = np.random.RandomState(0 ) lowerCamelCase__ : int = pipe( prompt=UpperCAmelCase , image=UpperCAmelCase , mask_image=UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=20 , generator=UpperCAmelCase , output_type='np' , ) lowerCamelCase__ : str = output.images lowerCamelCase__ : Tuple = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) lowerCamelCase__ : Dict = np.array([0.0_0_8_6, 0.0_0_7_7, 0.0_0_8_3, 0.0_0_9_3, 0.0_1_0_7, 0.0_1_3_9, 0.0_0_9_4, 0.0_0_9_7, 0.0_1_2_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
50
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
1
from __future__ import annotations from collections.abc import Iterator class lowerCAmelCase : def __init__( self : int , UpperCAmelCase : int ) -> None: lowerCamelCase__ : Dict = value lowerCamelCase__ : Node | None = None lowerCamelCase__ : Node | None = None class lowerCAmelCase : def __init__( self : Optional[int] , UpperCAmelCase : Node ) -> None: lowerCamelCase__ : Union[str, Any] = tree def A_ ( self : Union[str, Any] , UpperCAmelCase : Node | None ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : Tuple ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
50
import flax.linen as nn import jax import jax.numpy as jnp class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , UpperCAmelCase : Dict ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = hidden_states.shape lowerCamelCase__ : Union[str, Any] = jax.image.resize( UpperCAmelCase , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : str , UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = None UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels lowerCamelCase__ : Tuple = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : int = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Union[str, Any] = nn.Dense(UpperCAmelCase , dtype=self.dtype ) lowerCamelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : List[Any] = nn.Dropout(self.dropout_prob ) lowerCamelCase__ : Tuple = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Optional[Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCamelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCamelCase__ : Dict = nn.Conv( UpperCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=True ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = hidden_states lowerCamelCase__ : List[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[Any] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Any = self.conva(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.time_emb_proj(nn.swish(UpperCAmelCase ) ) lowerCamelCase__ : List[str] = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase , 1 ) , 1 ) lowerCamelCase__ : List[str] = hidden_states + temb lowerCamelCase__ : Optional[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[str] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.dropout(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : str = self.conva(UpperCAmelCase ) if self.conv_shortcut is not None: lowerCamelCase__ : Dict = self.conv_shortcut(UpperCAmelCase ) return hidden_states + residual
50
1
import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class lowerCAmelCase ( unittest.TestCase ): UpperCAmelCase__ = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING UpperCAmelCase__ = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def A_ ( self : Tuple , UpperCAmelCase : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> Any: lowerCamelCase__ : List[Any] = TextaTextGenerationPipeline(model=UpperCAmelCase , tokenizer=UpperCAmelCase ) return generator, ["Something to write", "Something else"] def A_ ( self : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Any ) -> str: lowerCamelCase__ : int = generator('Something there' ) self.assertEqual(UpperCAmelCase , [{'generated_text': ANY(UpperCAmelCase )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['generated_text'].startswith('Something there' ) ) lowerCamelCase__ : Dict = generator(['This is great !', 'Something else'] , num_return_sequences=2 , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ [{'generated_text': ANY(UpperCAmelCase )}, {'generated_text': ANY(UpperCAmelCase )}], [{'generated_text': ANY(UpperCAmelCase )}, {'generated_text': ANY(UpperCAmelCase )}], ] , ) lowerCamelCase__ : Optional[int] = generator( ['This is great !', 'Something else'] , num_return_sequences=2 , batch_size=2 , do_sample=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ [{'generated_text': ANY(UpperCAmelCase )}, {'generated_text': ANY(UpperCAmelCase )}], [{'generated_text': ANY(UpperCAmelCase )}, {'generated_text': ANY(UpperCAmelCase )}], ] , ) with self.assertRaises(UpperCAmelCase ): generator(4 ) @require_torch def A_ ( self : Tuple ) -> Dict: lowerCamelCase__ : Optional[int] = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='pt' ) # do_sample=False necessary for reproducibility lowerCamelCase__ : Optional[Any] = generator('Something there' , do_sample=UpperCAmelCase ) self.assertEqual(UpperCAmelCase , [{'generated_text': ''}] ) lowerCamelCase__ : int = 3 lowerCamelCase__ : Optional[Any] = generator( 'Something there' , num_return_sequences=UpperCAmelCase , num_beams=UpperCAmelCase , ) lowerCamelCase__ : Optional[Any] = [ {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': ''}, ] self.assertEqual(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = generator('This is a test' , do_sample=UpperCAmelCase , num_return_sequences=2 , return_tensors=UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ] , ) lowerCamelCase__ : int = generator.model.config.eos_token_id lowerCamelCase__ : Union[str, Any] = '<pad>' lowerCamelCase__ : Union[str, Any] = generator( ['This is a test', 'This is a second test'] , do_sample=UpperCAmelCase , num_return_sequences=2 , batch_size=2 , return_tensors=UpperCAmelCase , ) self.assertEqual( UpperCAmelCase , [ [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], ] , ) @require_tf def A_ ( self : List[Any] ) -> Optional[Any]: lowerCamelCase__ : Any = pipeline('text2text-generation' , model='patrickvonplaten/t5-tiny-random' , framework='tf' ) # do_sample=False necessary for reproducibility lowerCamelCase__ : str = generator('Something there' , do_sample=UpperCAmelCase ) self.assertEqual(UpperCAmelCase , [{'generated_text': ''}] )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Optional[Any] = set() # edges = list of graph's edges lowerCamelCase__ : List[str] = get_edges(_UpperCAmelCase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCamelCase__ , lowerCamelCase__ : str = edges.pop() chosen_vertices.add(_UpperCAmelCase ) chosen_vertices.add(_UpperCAmelCase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_UpperCAmelCase ) return chosen_vertices def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Union[str, Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
50
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : Optional[Any] = { """configuration_layoutlmv2""": ["""LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LayoutLMv2Config"""], """processing_layoutlmv2""": ["""LayoutLMv2Processor"""], """tokenization_layoutlmv2""": ["""LayoutLMv2Tokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = ["""LayoutLMv2TokenizerFast"""] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Union[str, Any] = ["""LayoutLMv2FeatureExtractor"""] _UpperCAmelCase : Any = ["""LayoutLMv2ImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ """LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """LayoutLMv2ForQuestionAnswering""", """LayoutLMv2ForSequenceClassification""", """LayoutLMv2ForTokenClassification""", """LayoutLMv2Layer""", """LayoutLMv2Model""", """LayoutLMv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Optional[Any] = { """configuration_upernet""": ["""UperNetConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ """UperNetForSemanticSegmentation""", """UperNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys _UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
1
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: if len(_UpperCAmelCase ) < k or k < 0: raise ValueError('Invalid Input' ) lowerCamelCase__ : Optional[int] = sum(array[:k] ) for i in range(len(_UpperCAmelCase ) - k ): lowerCamelCase__ : str = current_sum - array[i] + array[i + k] lowerCamelCase__ : List[str] = max(_UpperCAmelCase , _UpperCAmelCase ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() _UpperCAmelCase : List[Any] = [randint(-10_00, 10_00) for i in range(1_00)] _UpperCAmelCase : Optional[int] = randint(0, 1_10) print(F"""The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}""")
50
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : str=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : str=9 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=True , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : List[str]=5 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Union[str, Any]=37 , UpperCAmelCase : int=8 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=0.0_0_2 , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=0 , UpperCAmelCase : Union[str, Any]=0 , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , ) -> Union[str, Any]: lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = encoder_seq_length lowerCamelCase__ : int = decoder_seq_length # For common tests lowerCamelCase__ : List[str] = self.decoder_seq_length lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : List[Any] = use_attention_mask lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = d_ff lowerCamelCase__ : Optional[Any] = relative_attention_num_buckets lowerCamelCase__ : Any = dropout_rate lowerCamelCase__ : Any = initializer_factor lowerCamelCase__ : Union[str, Any] = eos_token_id lowerCamelCase__ : List[str] = pad_token_id lowerCamelCase__ : List[str] = decoder_start_token_id lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = decoder_layers def A_ ( self : List[Any] ) -> int: return TaConfig.from_pretrained('google/umt5-base' ) def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=None , ) -> List[str]: if attention_mask is None: lowerCamelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCamelCase__ : Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCamelCase__ : int = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase ) if decoder_head_mask is None: lowerCamelCase__ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A_ ( self : str ) -> List[str]: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCamelCase__ : List[str] = input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Union[str, Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Dict = self.get_config() lowerCamelCase__ : Tuple = config.num_attention_heads lowerCamelCase__ : Any = self.prepare_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, input_dict def A_ ( self : Tuple ) -> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Dict = self.prepare_config_and_inputs() return config, inputs_dict def A_ ( self : Optional[int] ) -> List[str]: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Union[str, Any] ) -> Dict: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Dict , ) -> str: lowerCamelCase__ : Dict = UMTaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , ) lowerCamelCase__ : Any = model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) lowerCamelCase__ : Dict = result.last_hidden_state lowerCamelCase__ : Any = result.past_key_values lowerCamelCase__ : List[Any] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[Any] , ) -> Optional[int]: lowerCamelCase__ : List[Any] = UMTaModel(config=UpperCAmelCase ).get_decoder().to(UpperCAmelCase ).eval() # first forward pass lowerCamelCase__ : Tuple = model(UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , use_cache=UpperCAmelCase ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) + 1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCamelCase__ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )['last_hidden_state'] lowerCamelCase__ : str = model(UpperCAmelCase , past_key_values=UpperCAmelCase )['last_hidden_state'] # select random slice lowerCamelCase__ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase__ : Tuple = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase__ : List[str] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase , UpperCAmelCase , atol=1e-3 ) ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = UMTaModel(config=UpperCAmelCase ).to(UpperCAmelCase ).half().eval() lowerCamelCase__ : Optional[int] = model(**UpperCAmelCase )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCAmelCase ).any().item() ) @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCAmelCase__ = (UMTaForConditionalGeneration,) if is_torch_available() else () UpperCAmelCase__ = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = True UpperCAmelCase__ = True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCAmelCase__ = [0.8, 0.9] def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Tuple = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCAmelCase , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: lowerCamelCase__ : int = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Any = config_and_inputs[0] lowerCamelCase__ : Any = UMTaForConditionalGeneration(UpperCAmelCase ).eval() model.to(UpperCAmelCase ) lowerCamelCase__ : Tuple = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), } for attn_name, (name, mask) in zip(UpperCAmelCase , head_masking.items() ): lowerCamelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCamelCase__ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ) lowerCamelCase__ : Tuple = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase , return_dict_in_generate=UpperCAmelCase , **UpperCAmelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCamelCase__ : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=UpperCAmelCase ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=UpperCAmelCase , legacy=UpperCAmelCase ) lowerCamelCase__ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowerCamelCase__ : Tuple = tokenizer(UpperCAmelCase , return_tensors='pt' , padding=UpperCAmelCase ).input_ids # fmt: off lowerCamelCase__ : Any = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model.generate(input_ids.to(UpperCAmelCase ) ) lowerCamelCase__ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowerCamelCase__ : Union[str, Any] = tokenizer.batch_decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
1
import os import sys _UpperCAmelCase : Optional[Any] = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) _UpperCAmelCase : int = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Dict: return AutoConfig.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> str: return AutoTokenizer.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Optional[int]: return AutoModel.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> int: return AutoModelForCausalLM.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> str: return AutoModelForMaskedLM.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Any: return AutoModelForSequenceClassification.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Optional[Any]: return AutoModelForQuestionAnswering.from_pretrained(*_UpperCAmelCase , **_UpperCAmelCase )
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCAmelCase ( __UpperCamelCase ): @slow @require_torch def A_ ( self : Tuple ) -> str: lowerCamelCase__ : Optional[int] = EncoderDecoderModel.from_encoder_decoder_pretrained('prajjwal1/bert-tiny' , 'prajjwal1/bert-tiny' ) lowerCamelCase__ : List[Any] = BertTokenizer.from_pretrained('bert-base-uncased' ) lowerCamelCase__ : Tuple = bertabert.config.encoder.vocab_size lowerCamelCase__ : Optional[int] = tokenizer.sep_token_id lowerCamelCase__ : List[Any] = tokenizer.cls_token_id lowerCamelCase__ : str = 128 lowerCamelCase__ : Optional[Any] = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='train[:1%]' ) lowerCamelCase__ : Any = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='validation[:1%]' ) lowerCamelCase__ : List[Any] = train_dataset.select(range(32 ) ) lowerCamelCase__ : Dict = val_dataset.select(range(16 ) ) lowerCamelCase__ : str = 4 def _map_to_encoder_decoder_inputs(UpperCAmelCase : int ): # Tokenizer will automatically set [BOS] <text> [EOS] lowerCamelCase__ : List[str] = tokenizer(batch['article'] , padding='max_length' , truncation=UpperCAmelCase , max_length=512 ) lowerCamelCase__ : Optional[Any] = tokenizer(batch['highlights'] , padding='max_length' , truncation=UpperCAmelCase , max_length=128 ) lowerCamelCase__ : int = inputs.input_ids lowerCamelCase__ : List[str] = inputs.attention_mask lowerCamelCase__ : Optional[int] = outputs.input_ids lowerCamelCase__ : Dict = outputs.input_ids.copy() lowerCamelCase__ : Union[str, Any] = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['labels'] ] lowerCamelCase__ : Union[str, Any] = outputs.attention_mask assert all(len(UpperCAmelCase ) == 512 for x in inputs.input_ids ) assert all(len(UpperCAmelCase ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(UpperCAmelCase : Optional[int] ): lowerCamelCase__ : int = pred.label_ids lowerCamelCase__ : Optional[int] = pred.predictions # all unnecessary tokens are removed lowerCamelCase__ : Optional[int] = tokenizer.batch_decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.batch_decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase ) lowerCamelCase__ : Any = sum([int(pred_str[i] == label_str[i] ) for i in range(len(UpperCAmelCase ) )] ) / len(UpperCAmelCase ) return {"accuracy": accuracy} # map train dataset lowerCamelCase__ : int = train_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCAmelCase , batch_size=UpperCAmelCase , remove_columns=['article', 'highlights'] , ) train_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) # same for validation dataset lowerCamelCase__ : Optional[int] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=UpperCAmelCase , batch_size=UpperCAmelCase , remove_columns=['article', 'highlights'] , ) val_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) lowerCamelCase__ : Dict = self.get_auto_remove_tmp_dir() lowerCamelCase__ : List[str] = SeqaSeqTrainingArguments( output_dir=UpperCAmelCase , per_device_train_batch_size=UpperCAmelCase , per_device_eval_batch_size=UpperCAmelCase , predict_with_generate=UpperCAmelCase , evaluation_strategy='steps' , do_train=UpperCAmelCase , do_eval=UpperCAmelCase , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer lowerCamelCase__ : List[str] = SeqaSeqTrainer( model=UpperCAmelCase , args=UpperCAmelCase , compute_metrics=_compute_metrics , train_dataset=UpperCAmelCase , eval_dataset=UpperCAmelCase , tokenizer=UpperCAmelCase , ) # start training trainer.train()
50
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase ): @register_to_config def __init__( self : List[str] , UpperCAmelCase : int = 65536 , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 0 , UpperCAmelCase : str = "fourier" , UpperCAmelCase : bool = True , UpperCAmelCase : bool = False , UpperCAmelCase : float = 0.0 , UpperCAmelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase : str = None , UpperCAmelCase : Tuple[int] = (32, 32, 64) , UpperCAmelCase : str = None , UpperCAmelCase : int = 8 , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = False , ) -> List[Any]: super().__init__() lowerCamelCase__ : Optional[int] = sample_size # time if time_embedding_type == "fourier": lowerCamelCase__ : Optional[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase , log=UpperCAmelCase , flip_sin_to_cos=UpperCAmelCase ) lowerCamelCase__ : Any = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCamelCase__ : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase , downscale_freq_shift=UpperCAmelCase ) lowerCamelCase__ : Dict = block_out_channels[0] if use_timestep_embedding: lowerCamelCase__ : str = block_out_channels[0] * 4 lowerCamelCase__ : List[Any] = TimestepEmbedding( in_channels=UpperCAmelCase , time_embed_dim=UpperCAmelCase , act_fn=UpperCAmelCase , out_dim=block_out_channels[0] , ) lowerCamelCase__ : Any = nn.ModuleList([] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Optional[int] = None # down lowerCamelCase__ : Optional[int] = in_channels for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = output_channel lowerCamelCase__ : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Optional[int] = get_down_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Optional[int] = get_mid_block( UpperCAmelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase , add_downsample=UpperCAmelCase , ) # up lowerCamelCase__ : Optional[int] = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = reversed_block_out_channels[0] if out_block_type is None: lowerCamelCase__ : List[str] = out_channels else: lowerCamelCase__ : Any = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Tuple = output_channel lowerCamelCase__ : Union[str, Any] = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase ) - 1 else final_upsample_channels ) lowerCamelCase__ : List[str] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Dict = get_up_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : int = output_channel # out lowerCamelCase__ : int = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCamelCase__ : List[Any] = get_out_block( out_block_type=UpperCAmelCase , num_groups_out=UpperCAmelCase , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase , act_fn=UpperCAmelCase , fc_dim=block_out_channels[-1] // 4 , ) def A_ ( self : List[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Union[torch.Tensor, float, int] , UpperCAmelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: lowerCamelCase__ : Optional[Any] = timestep if not torch.is_tensor(UpperCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(sample.device ) lowerCamelCase__ : Optional[int] = self.time_proj(UpperCAmelCase ) if self.config.use_timestep_embedding: lowerCamelCase__ : str = self.time_mlp(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = timestep_embed[..., None] lowerCamelCase__ : str = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCamelCase__ : str = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCamelCase__ : str = () for downsample_block in self.down_blocks: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = downsample_block(hidden_states=UpperCAmelCase , temb=UpperCAmelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCamelCase__ : Optional[Any] = self.mid_block(UpperCAmelCase , UpperCAmelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCamelCase__ : Dict = down_block_res_samples[-1:] lowerCamelCase__ : Optional[Any] = down_block_res_samples[:-1] lowerCamelCase__ : Any = upsample_block(UpperCAmelCase , res_hidden_states_tuple=UpperCAmelCase , temb=UpperCAmelCase ) # 5. post-process if self.out_block: lowerCamelCase__ : Any = self.out_block(UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase )
50
1
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class lowerCAmelCase ( unittest.TestCase ): def A_ ( self : int ) -> Optional[Any]: debug_launcher(test_script.main ) def A_ ( self : List[str] ) -> Optional[int]: debug_launcher(test_ops.main )
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCAmelCase : Optional[Any] = { """configuration_convbert""": ["""CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvBertConfig""", """ConvBertOnnxConfig"""], """tokenization_convbert""": ["""ConvBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = ["""ConvBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ """CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvBertForMaskedLM""", """ConvBertForMultipleChoice""", """ConvBertForQuestionAnswering""", """ConvBertForSequenceClassification""", """ConvBertForTokenClassification""", """ConvBertLayer""", """ConvBertModel""", """ConvBertPreTrainedModel""", """load_tf_weights_in_convbert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ """TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFConvBertForMaskedLM""", """TFConvBertForMultipleChoice""", """TFConvBertForQuestionAnswering""", """TFConvBertForSequenceClassification""", """TFConvBertForTokenClassification""", """TFConvBertLayer""", """TFConvBertModel""", """TFConvBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[int] = [] embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", F"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", F"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", F"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", F"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Tuple = [] attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", F"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", F"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", F"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", F"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] token.append((F"""cvt.encoder.stages.{idx}.cls_token""", 'stage2.cls_token') ) return token def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Tuple = 'imagenet-1k-id2label.json' lowerCamelCase__ : Union[str, Any] = 1000 lowerCamelCase__ : Optional[Any] = 'huggingface/label-files' lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowerCamelCase__ : int = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : List[str] = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowerCamelCase__ : List[Any] = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowerCamelCase__ : Dict = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowerCamelCase__ : Optional[Any] = [2, 2, 20] lowerCamelCase__ : Optional[int] = [3, 12, 16] lowerCamelCase__ : str = [192, 768, 1024] lowerCamelCase__ : Any = CvtForImageClassification(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowerCamelCase__ : Optional[int] = OrderedDict() lowerCamelCase__ : Tuple = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowerCamelCase__ : Optional[Any] = list_of_state_dict + cls_token(_UpperCAmelCase ) lowerCamelCase__ : str = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowerCamelCase__ : str = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : int = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : str = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( """--cvt_model""", default="""cvt-w24""", type=str, help="""Name of the cvt model you'd like to convert.""", ) parser.add_argument( """--image_size""", default=3_84, type=int, help="""Input Image Size""", ) parser.add_argument( """--cvt_file_name""", default=R"""cvtmodels\CvT-w24-384x384-IN-22k.pth""", type=str, help="""Input Image Size""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
50
1
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: return 1 / (1 + np.exp(-z )) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ : List[Any] = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=7_0000 ) -> List[str]: lowerCamelCase__ : str = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): lowerCamelCase__ : List[str] = np.dot(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = sigmoid_function(_UpperCAmelCase ) lowerCamelCase__ : Tuple = np.dot(x.T , h - y ) / y.size lowerCamelCase__ : List[Any] = theta - alpha * gradient # updating the weights lowerCamelCase__ : List[str] = np.dot(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : List[Any] = sigmoid_function(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 100 == 0: print(F"""loss: {j} \t""" ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = datasets.load_iris() _UpperCAmelCase : Optional[int] = iris.data[:, :2] _UpperCAmelCase : Tuple = (iris.target != 0) * 1 _UpperCAmelCase : List[Any] = 0.1 _UpperCAmelCase : str = logistic_reg(alpha, x, y, max_iterations=7_00_00) print("""theta: """, theta) # printing the theta i.e our weights vector def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color="""b""", label="""0""") plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color="""r""", label="""1""") ((_UpperCAmelCase) ,(_UpperCAmelCase)) : str = (x[:, 0].min(), x[:, 0].max()) ((_UpperCAmelCase) ,(_UpperCAmelCase)) : Optional[int] = (x[:, 1].min(), x[:, 1].max()) ((_UpperCAmelCase) ,(_UpperCAmelCase)) : Union[str, Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) _UpperCAmelCase : Any = np.c_[xxa.ravel(), xxa.ravel()] _UpperCAmelCase : List[str] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors="""black""") plt.legend() plt.show()
50
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
1
import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Optional[Any] = XCLIPTextConfig() # derive patch size from model name lowerCamelCase__ : Optional[Any] = model_name.find('patch' ) lowerCamelCase__ : Tuple = int(model_name[start_idx + len('patch' ) : start_idx + len('patch' ) + 2] ) lowerCamelCase__ : str = XCLIPVisionConfig(patch_size=_UpperCAmelCase , num_frames=_UpperCAmelCase ) if "large" in model_name: lowerCamelCase__ : int = 768 lowerCamelCase__ : Dict = 3072 lowerCamelCase__ : Optional[Any] = 12 lowerCamelCase__ : List[str] = 1024 lowerCamelCase__ : str = 4096 lowerCamelCase__ : Dict = 16 lowerCamelCase__ : int = 24 lowerCamelCase__ : List[Any] = 768 lowerCamelCase__ : int = 3072 if model_name == "xclip-large-patch14-16-frames": lowerCamelCase__ : Optional[Any] = 336 lowerCamelCase__ : List[Any] = XCLIPConfig.from_text_vision_configs(_UpperCAmelCase , _UpperCAmelCase ) if "large" in model_name: lowerCamelCase__ : List[str] = 768 return config def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: # text encoder if name == "token_embedding.weight": lowerCamelCase__ : int = name.replace('token_embedding.weight' , 'text_model.embeddings.token_embedding.weight' ) if name == "positional_embedding": lowerCamelCase__ : List[Any] = name.replace('positional_embedding' , 'text_model.embeddings.position_embedding.weight' ) if "ln_1" in name: lowerCamelCase__ : Dict = name.replace('ln_1' , 'layer_norm1' ) if "ln_2" in name: lowerCamelCase__ : Any = name.replace('ln_2' , 'layer_norm2' ) if "c_fc" in name: lowerCamelCase__ : str = name.replace('c_fc' , 'fc1' ) if "c_proj" in name: lowerCamelCase__ : int = name.replace('c_proj' , 'fc2' ) if name.startswith('transformer.resblocks' ): lowerCamelCase__ : Any = name.replace('transformer.resblocks' , 'text_model.encoder.layers' ) if "attn.out_proj" in name and "message" not in name: lowerCamelCase__ : Union[str, Any] = name.replace('attn.out_proj' , 'self_attn.out_proj' ) if "ln_final" in name: lowerCamelCase__ : List[Any] = name.replace('ln_final' , 'text_model.final_layer_norm' ) # visual encoder if name == "visual.class_embedding": lowerCamelCase__ : List[str] = name.replace('visual.class_embedding' , 'vision_model.embeddings.class_embedding' ) if name == "visual.positional_embedding": lowerCamelCase__ : Tuple = name.replace('visual.positional_embedding' , 'vision_model.embeddings.position_embedding.weight' ) if name.startswith('visual.transformer.resblocks' ): lowerCamelCase__ : List[Any] = name.replace('visual.transformer.resblocks' , 'vision_model.encoder.layers' ) if "visual.conv1" in name: lowerCamelCase__ : str = name.replace('visual.conv1' , 'vision_model.embeddings.patch_embedding' ) if "visual.ln_pre" in name: lowerCamelCase__ : Tuple = name.replace('visual.ln_pre' , 'vision_model.pre_layernorm' ) if "visual.ln_post" in name: lowerCamelCase__ : List[str] = name.replace('visual.ln_post' , 'vision_model.post_layernorm' ) if "visual.proj" in name: lowerCamelCase__ : str = name.replace('visual.proj' , 'visual_projection.weight' ) if "text_projection" in name: lowerCamelCase__ : Optional[Any] = name.replace('text_projection' , 'text_projection.weight' ) # things on top if "prompts_visual_proj" in name: lowerCamelCase__ : Union[str, Any] = name.replace('prompts_visual_proj' , 'prompts_visual_projection' ) if "prompts_visual_ln" in name: lowerCamelCase__ : Any = name.replace('prompts_visual_ln' , 'prompts_visual_layernorm' ) # mit if name == "mit.positional_embedding": lowerCamelCase__ : Any = name.replace('positional' , 'position' ) if name.startswith('mit.resblocks' ): lowerCamelCase__ : Optional[Any] = name.replace('mit.resblocks' , 'mit.encoder.layers' ) # prompts generator if name.startswith('prompts_generator.norm' ): lowerCamelCase__ : Tuple = name.replace('prompts_generator.norm' , 'prompts_generator.layernorm' ) return name def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Tuple = orig_state_dict.pop(_UpperCAmelCase ) if "attn.in_proj" in key: lowerCamelCase__ : Union[str, Any] = key.split('.' ) if key.startswith('visual' ): lowerCamelCase__ : Optional[Any] = key_split[3] lowerCamelCase__ : Tuple = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: lowerCamelCase__ : Optional[int] = val[ :dim, : ] lowerCamelCase__ : Optional[Any] = val[ dim : dim * 2, : ] lowerCamelCase__ : int = val[ -dim:, : ] else: lowerCamelCase__ : List[Any] = val[ :dim ] lowerCamelCase__ : Optional[Any] = val[ dim : dim * 2 ] lowerCamelCase__ : Optional[int] = val[ -dim: ] else: if "weight" in key: lowerCamelCase__ : int = val[ :dim, : ] lowerCamelCase__ : Dict = val[ dim : dim * 2, : ] lowerCamelCase__ : int = val[ -dim:, : ] else: lowerCamelCase__ : Optional[int] = val[:dim] lowerCamelCase__ : str = val[ dim : dim * 2 ] lowerCamelCase__ : Optional[Any] = val[-dim:] elif key.startswith('mit' ): lowerCamelCase__ : List[str] = key_split[2] lowerCamelCase__ : int = config.vision_config.mit_hidden_size if "weight" in key: lowerCamelCase__ : List[str] = val[:dim, :] lowerCamelCase__ : str = val[dim : dim * 2, :] lowerCamelCase__ : Union[str, Any] = val[-dim:, :] else: lowerCamelCase__ : Optional[Any] = val[:dim] lowerCamelCase__ : List[str] = val[dim : dim * 2] lowerCamelCase__ : Any = val[-dim:] else: lowerCamelCase__ : Optional[int] = key_split[2] lowerCamelCase__ : Dict = config.text_config.hidden_size if "weight" in key: lowerCamelCase__ : Optional[int] = val[:dim, :] lowerCamelCase__ : int = val[ dim : dim * 2, : ] lowerCamelCase__ : Any = val[-dim:, :] else: lowerCamelCase__ : Any = val[:dim] lowerCamelCase__ : List[Any] = val[ dim : dim * 2 ] lowerCamelCase__ : str = val[-dim:] else: lowerCamelCase__ : Optional[int] = rename_key(_UpperCAmelCase ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: lowerCamelCase__ : List[Any] = val.T lowerCamelCase__ : Optional[Any] = val return orig_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: if num_frames == 8: lowerCamelCase__ : Any = 'eating_spaghetti_8_frames.npy' elif num_frames == 16: lowerCamelCase__ : Union[str, Any] = 'eating_spaghetti.npy' elif num_frames == 32: lowerCamelCase__ : str = 'eating_spaghetti_32_frames.npy' lowerCamelCase__ : Dict = hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename=_UpperCAmelCase , repo_type='dataset' , ) lowerCamelCase__ : str = np.load(_UpperCAmelCase ) return list(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=False ) -> Union[str, Any]: lowerCamelCase__ : Any = { # fully supervised kinetics-400 checkpoints 'xclip-base-patch32': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth', 'xclip-base-patch32-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth' ), 'xclip-base-patch16': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth', 'xclip-base-patch16-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth' ), 'xclip-large-patch14': 'https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb', 'xclip-large-patch14-16-frames': 'https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f', # fully supervised kinetics-600 checkpoints 'xclip-base-patch16-kinetics-600': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth' ), 'xclip-base-patch16-kinetics-600-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth' ), 'xclip-large-patch14-kinetics-600': 'https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be', # few shot 'xclip-base-patch16-hmdb-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth' ), 'xclip-base-patch16-hmdb-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth' ), 'xclip-base-patch16-hmdb-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth' ), 'xclip-base-patch16-hmdb-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth' ), 'xclip-base-patch16-ucf-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth' ), 'xclip-base-patch16-ucf-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth' ), 'xclip-base-patch16-ucf-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth' ), 'xclip-base-patch16-ucf-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth' ), # zero shot 'xclip-base-patch16-zero-shot': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth', } lowerCamelCase__ : List[Any] = model_to_url[model_name] lowerCamelCase__ : Optional[Any] = 8 if "16-frames" in model_name: lowerCamelCase__ : Union[str, Any] = 16 elif "shot" in model_name: lowerCamelCase__ : Optional[int] = 32 lowerCamelCase__ : Tuple = get_xclip_config(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : List[Any] = XCLIPModel(_UpperCAmelCase ) model.eval() if "drive" in checkpoint_url: lowerCamelCase__ : Dict = 'pytorch_model.bin' gdown.cached_download(_UpperCAmelCase , _UpperCAmelCase , quiet=_UpperCAmelCase ) lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location='cpu' )['model'] else: lowerCamelCase__ : Tuple = torch.hub.load_state_dict_from_url(_UpperCAmelCase )['model'] lowerCamelCase__ : Any = convert_state_dict(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : List[Any] = XCLIPModel(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() lowerCamelCase__ : Any = 336 if model_name == 'xclip-large-patch14-16-frames' else 224 lowerCamelCase__ : List[str] = VideoMAEImageProcessor(size=_UpperCAmelCase ) lowerCamelCase__ : str = CLIPTokenizer.from_pretrained('openai/clip-vit-base-patch32' ) lowerCamelCase__ : List[Any] = CLIPTokenizerFast.from_pretrained('openai/clip-vit-base-patch32' ) lowerCamelCase__ : Optional[int] = XCLIPProcessor(image_processor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) lowerCamelCase__ : int = prepare_video(_UpperCAmelCase ) lowerCamelCase__ : Tuple = processor( text=['playing sports', 'eating spaghetti', 'go shopping'] , videos=_UpperCAmelCase , return_tensors='pt' , padding=_UpperCAmelCase ) print('Shape of pixel values:' , inputs.pixel_values.shape ) with torch.no_grad(): lowerCamelCase__ : List[Any] = model(**_UpperCAmelCase ) # Verify outputs lowerCamelCase__ : str = outputs.logits_per_video lowerCamelCase__ : Union[str, Any] = logits_per_video.softmax(dim=1 ) print('Probs:' , _UpperCAmelCase ) # kinetics-400 if model_name == "xclip-base-patch32": lowerCamelCase__ : Optional[Any] = torch.tensor([[0.0_019, 0.9_951, 0.0_030]] ) elif model_name == "xclip-base-patch32-16-frames": lowerCamelCase__ : Union[str, Any] = torch.tensor([[7.0_999e-04, 9.9_883e-01, 4.5_580e-04]] ) elif model_name == "xclip-base-patch16": lowerCamelCase__ : int = torch.tensor([[0.0_083, 0.9_681, 0.0_236]] ) elif model_name == "xclip-base-patch16-16-frames": lowerCamelCase__ : Any = torch.tensor([[7.6_937e-04, 9.9_728e-01, 1.9_473e-03]] ) elif model_name == "xclip-large-patch14": lowerCamelCase__ : Tuple = torch.tensor([[0.0_062, 0.9_864, 0.0_075]] ) elif model_name == "xclip-large-patch14-16-frames": lowerCamelCase__ : Optional[int] = torch.tensor([[3.3_877e-04, 9.9_937e-01, 2.8_888e-04]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": lowerCamelCase__ : Tuple = torch.tensor([[0.0_555, 0.8_914, 0.0_531]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": lowerCamelCase__ : Tuple = torch.tensor([[3.8_554e-04, 9.9_929e-01, 3.2_754e-04]] ) elif model_name == "xclip-large-patch14-kinetics-600": lowerCamelCase__ : List[Any] = torch.tensor([[0.0_036, 0.9_920, 0.0_045]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": lowerCamelCase__ : Tuple = torch.tensor([[7.1_890e-06, 9.9_994e-01, 5.6_559e-05]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": lowerCamelCase__ : Any = torch.tensor([[1.0_320e-05, 9.9_993e-01, 6.2_435e-05]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": lowerCamelCase__ : Optional[Any] = torch.tensor([[4.1_377e-06, 9.9_990e-01, 9.8_386e-05]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": lowerCamelCase__ : Optional[int] = torch.tensor([[4.1_347e-05, 9.9_962e-01, 3.3_411e-04]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": lowerCamelCase__ : int = torch.tensor([[8.5_857e-05, 9.9_928e-01, 6.3_291e-04]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": lowerCamelCase__ : Dict = torch.tensor([[8.5_857e-05, 9.9_928e-01, 6.3_291e-04]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": lowerCamelCase__ : List[Any] = torch.tensor([[0.0_027, 0.9_904, 0.0_070]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": lowerCamelCase__ : Union[str, Any] = torch.tensor([[9.8_219e-04, 9.9_593e-01, 3.0_863e-03]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": lowerCamelCase__ : int = torch.tensor([[3.5_082e-04, 9.9_785e-01, 1.7_966e-03]] ) else: raise ValueError(F"""Model name {model_name} not supported""" ) assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) if push_to_hub: print('Pushing model, processor and slow tokenizer files to the hub...' ) model.push_to_hub(_UpperCAmelCase , organization='nielsr' ) processor.push_to_hub(_UpperCAmelCase , organization='nielsr' ) slow_tokenizer.push_to_hub(_UpperCAmelCase , organization='nielsr' ) if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""xclip-base-patch32""", type=str, help="""Name of the model.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } _UpperCAmelCase : Any = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]="[UNK]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Tuple="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=None , **UpperCAmelCase : Optional[int] , ) -> str: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : str = strip_accents lowerCamelCase__ : Optional[Any] = tokenize_chinese_chars lowerCamelCase__ : int = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : str = do_lower_case def A_ ( self : Optional[int] , UpperCAmelCase : int , **UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH lowerCamelCase__ : Optional[int] = text lowerCamelCase__ : Dict = kwargs.pop('text_pair' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = kwargs.pop('return_tensors' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: lowerCamelCase__ : Tuple = batch_text_pair[idx] else: lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Any = encoded_candidates.get('input_ids' ) lowerCamelCase__ : Union[str, Any] = encoded_candidates.get('attention_mask' ) lowerCamelCase__ : Tuple = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) lowerCamelCase__ : int = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ) -> List[str]: lowerCamelCase__ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = { """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """lilt""" def __init__( self : Dict , UpperCAmelCase : int=30522 , UpperCAmelCase : Tuple=768 , UpperCAmelCase : Optional[int]=12 , UpperCAmelCase : List[str]=12 , UpperCAmelCase : Union[str, Any]=3072 , UpperCAmelCase : Tuple="gelu" , UpperCAmelCase : Optional[int]=0.1 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=512 , UpperCAmelCase : List[str]=2 , UpperCAmelCase : str=0.0_2 , UpperCAmelCase : List[Any]=1e-12 , UpperCAmelCase : List[str]=0 , UpperCAmelCase : Union[str, Any]="absolute" , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=4 , UpperCAmelCase : Union[str, Any]=1024 , **UpperCAmelCase : str , ) -> Dict: super().__init__(pad_token_id=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Dict = hidden_size lowerCamelCase__ : Tuple = num_hidden_layers lowerCamelCase__ : Optional[Any] = num_attention_heads lowerCamelCase__ : Tuple = hidden_act lowerCamelCase__ : Union[str, Any] = intermediate_size lowerCamelCase__ : List[Any] = hidden_dropout_prob lowerCamelCase__ : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase__ : Optional[int] = max_position_embeddings lowerCamelCase__ : Tuple = type_vocab_size lowerCamelCase__ : Dict = initializer_range lowerCamelCase__ : Union[str, Any] = layer_norm_eps lowerCamelCase__ : Any = position_embedding_type lowerCamelCase__ : List[str] = classifier_dropout lowerCamelCase__ : int = channel_shrink_ratio lowerCamelCase__ : int = max_ad_position_embeddings
50
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _UpperCAmelCase : Optional[Any] = logging.getLogger(__name__) @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if is_torch_available(): import torch from torch.utils.data import Dataset class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 def __init__( self : int , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : List[str]=False , UpperCAmelCase : bool = False , ) -> List[str]: lowerCamelCase__ : int = hans_processors[task]() lowerCamelCase__ : Optional[Any] = os.path.join( UpperCAmelCase , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase ) , UpperCAmelCase , ) , ) lowerCamelCase__ : int = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = label_list[2], label_list[1] lowerCamelCase__ : List[str] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : str = cached_features_file + '.lock' with FileLock(UpperCAmelCase ): if os.path.exists(UpperCAmelCase ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) lowerCamelCase__ : int = torch.load(UpperCAmelCase ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) lowerCamelCase__ : str = ( processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) ) logger.info('Training examples: %s' , len(UpperCAmelCase ) ) lowerCamelCase__ : Dict = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) logger.info('Saving features into cached file %s' , UpperCAmelCase ) torch.save(self.features , UpperCAmelCase ) def __len__( self : Optional[int] ) -> Optional[Any]: return len(self.features ) def __getitem__( self : Tuple , UpperCAmelCase : Dict ) -> InputFeatures: return self.features[i] def A_ ( self : int ) -> int: return self.label_list if is_tf_available(): import tensorflow as tf class lowerCAmelCase : UpperCAmelCase__ = 42 def __init__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = 128 , UpperCAmelCase : Any=False , UpperCAmelCase : bool = False , ) -> Union[str, Any]: lowerCamelCase__ : Any = hans_processors[task]() lowerCamelCase__ : Optional[Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : str = label_list[2], label_list[1] lowerCamelCase__ : Optional[int] = label_list lowerCamelCase__ : int = processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCamelCase__ : Optional[int] = tf.data.Dataset.from_generator( UpperCAmelCase , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def A_ ( self : Any ) -> Any: return self.dataset def __len__( self : Tuple ) -> int: return len(self.features ) def __getitem__( self : List[str] , UpperCAmelCase : Any ) -> InputFeatures: return self.features[i] def A_ ( self : Dict ) -> str: return self.label_list class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : int , UpperCAmelCase : List[Any] ) -> int: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_train_set.txt' ) ) , 'train' ) def A_ ( self : Any , UpperCAmelCase : int ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def A_ ( self : Any ) -> List[Any]: return ["contradiction", "entailment", "neutral"] def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> List[str]: lowerCamelCase__ : List[str] = [] for i, line in enumerate(UpperCAmelCase ): if i == 0: continue lowerCamelCase__ : Tuple = '%s-%s' % (set_type, line[0]) lowerCamelCase__ : str = line[5] lowerCamelCase__ : Dict = line[6] lowerCamelCase__ : int = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCamelCase__ : Dict = line[0] examples.append(InputExample(guid=UpperCAmelCase , text_a=UpperCAmelCase , text_b=UpperCAmelCase , label=UpperCAmelCase , pairID=UpperCAmelCase ) ) return examples def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> Optional[int]: lowerCamelCase__ : int = {label: i for i, label in enumerate(_UpperCAmelCase )} lowerCamelCase__ : List[Any] = [] for ex_index, example in tqdm.tqdm(enumerate(_UpperCAmelCase ) , desc='convert examples to features' ): if ex_index % 1_0000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCamelCase__ : List[Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=_UpperCAmelCase , max_length=_UpperCAmelCase , padding='max_length' , truncation=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , ) lowerCamelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCamelCase__ : Optional[int] = int(example.pairID ) features.append(InputFeatures(**_UpperCAmelCase , label=_UpperCAmelCase , pairID=_UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features _UpperCAmelCase : str = { """hans""": 3, } _UpperCAmelCase : List[Any] = { """hans""": HansProcessor, }
50
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _UpperCAmelCase : Optional[int] = 16 _UpperCAmelCase : int = 32 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase = 16 ) -> Dict: lowerCamelCase__ : Dict = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCamelCase__ : List[Any] = load_dataset('glue' , 'mrpc' ) def tokenize_function(_UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase__ : List[str] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase__ : Any = datasets.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase__ : str = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase__ : List[Any] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase__ : Optional[int] = 16 elif accelerator.mixed_precision != "no": lowerCamelCase__ : Tuple = 8 else: lowerCamelCase__ : Tuple = None return tokenizer.pad( _UpperCAmelCase , padding='longest' , max_length=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_tensors='pt' , ) # Instantiate dataloaders. lowerCamelCase__ : str = DataLoader( tokenized_datasets['train'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = DataLoader( tokenized_datasets['validation'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _UpperCAmelCase : str = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , _UpperCAmelCase ) == "1": lowerCamelCase__ : Dict = 2 # New Code # lowerCamelCase__ : str = int(args.gradient_accumulation_steps ) lowerCamelCase__ : int = int(args.local_sgd_steps ) # Initialize accelerator lowerCamelCase__ : str = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_UpperCAmelCase ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase__ : Any = config['lr'] lowerCamelCase__ : int = int(config['num_epochs'] ) lowerCamelCase__ : List[str] = int(config['seed'] ) lowerCamelCase__ : Any = int(config['batch_size'] ) lowerCamelCase__ : Dict = evaluate.load('glue' , 'mrpc' ) set_seed(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = get_dataloaders(_UpperCAmelCase , _UpperCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase__ : Tuple = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=_UpperCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase__ : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase__ : Optional[int] = AdamW(params=model.parameters() , lr=_UpperCAmelCase ) # Instantiate scheduler lowerCamelCase__ : Tuple = get_linear_schedule_with_warmup( optimizer=_UpperCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_UpperCAmelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = accelerator.prepare( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Now we train the model for epoch in range(_UpperCAmelCase ): model.train() with LocalSGD( accelerator=_UpperCAmelCase , model=_UpperCAmelCase , local_sgd_steps=_UpperCAmelCase , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_UpperCAmelCase ): lowerCamelCase__ : str = model(**_UpperCAmelCase ) lowerCamelCase__ : List[str] = output.loss accelerator.backward(_UpperCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**_UpperCAmelCase ) lowerCamelCase__ : Dict = outputs.logits.argmax(dim=-1 ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=_UpperCAmelCase , references=_UpperCAmelCase , ) lowerCamelCase__ : Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : Optional[Any] = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=_UpperCAmelCase , default=_UpperCAmelCase , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=_UpperCAmelCase , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument( '--local_sgd_steps' , type=_UpperCAmelCase , default=8 , help='Number of local SGD steps or None to disable local SGD' ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) lowerCamelCase__ : Optional[Any] = parser.parse_args() lowerCamelCase__ : Optional[Any] = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
50
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[Any] = """▁""" _UpperCAmelCase : Optional[Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = BertGenerationTokenizer UpperCAmelCase__ = False UpperCAmelCase__ = True def A_ ( self : List[Any] ) -> List[str]: super().setUp() lowerCamelCase__ : Dict = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : List[str] = '<s>' lowerCamelCase__ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : List[str] ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(UpperCAmelCase ) , 1002 ) def A_ ( self : List[Any] ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowerCamelCase__ : List[str] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowerCamelCase__ : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def A_ ( self : Dict ) -> Tuple: return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def A_ ( self : Optional[int] ) -> List[str]: lowerCamelCase__ : Union[str, Any] = 'Hello World!' lowerCamelCase__ : Dict = [18536, 2260, 101] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @slow def A_ ( self : Optional[Any] ) -> str: lowerCamelCase__ : List[Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCamelCase__ : Any = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @require_torch @slow def A_ ( self : int ) -> Optional[Any]: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowerCamelCase__ : str = list(self.big_tokenizer.get_vocab().keys() )[:10] lowerCamelCase__ : int = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.big_tokenizer.encode_plus(UpperCAmelCase , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Tuple = BertGenerationConfig() lowerCamelCase__ : Optional[Any] = BertGenerationEncoder(UpperCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase ) model(**UpperCAmelCase ) @slow def A_ ( self : Optional[int] ) -> List[Any]: # fmt: off lowerCamelCase__ : Any = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 6008_5147_5143 ) -> int: try: lowerCamelCase__ : Optional[Any] = int(_UpperCAmelCase ) except (TypeError, ValueError): raise TypeError('Parameter n must be int or castable to int.' ) if n <= 0: raise ValueError('Parameter n must be greater than or equal to one.' ) lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : str = 2 while i * i <= n: while n % i == 0: lowerCamelCase__ : Tuple = i n //= i i += 1 if n > 1: lowerCamelCase__ : Tuple = n return int(_UpperCAmelCase ) if __name__ == "__main__": print(F"""{solution() = }""")
50
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _UpperCAmelCase : str = pytest.mark.integration @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : List[Any] ) -> Union[str, Any]: lowerCamelCase__ : List[Any] = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(UpperCAmelCase ) for x in np.arange(30 ).tolist()]} ) return dset def A_ ( self : Optional[Any] ) -> Optional[int]: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() lowerCamelCase__ : List[Any] = dset.map( lambda UpperCAmelCase , UpperCAmelCase : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=UpperCAmelCase , keep_in_memory=UpperCAmelCase ) lowerCamelCase__ : Tuple = dset.add_faiss_index('vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def A_ ( self : Union[str, Any] ) -> int: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : List[str] ) -> Tuple: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : Any ) -> Optional[Any]: lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(UpperCAmelCase , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def A_ ( self : Dict ) -> Dict: from elasticsearch import Elasticsearch lowerCamelCase__ : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : List[Any] = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) lowerCamelCase__ : int = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} lowerCamelCase__ : List[str] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Any ) -> Dict: import faiss lowerCamelCase__ : Tuple = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query lowerCamelCase__ : int = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Any = 1 lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = index.search(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries lowerCamelCase__ : str = np.eye(5 , dtype=np.floataa )[::-1] lowerCamelCase__ , lowerCamelCase__ : Dict = index.search_batch(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search_batch , queries[0] ) lowerCamelCase__ : str = [scores[0] for scores in total_scores] lowerCamelCase__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , UpperCAmelCase ) def A_ ( self : List[Any] ) -> List[Any]: import faiss lowerCamelCase__ : Any = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) lowerCamelCase__ : Tuple = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(UpperCAmelCase ): lowerCamelCase__ : List[str] = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def A_ ( self : List[str] ) -> Optional[int]: import faiss lowerCamelCase__ : Optional[Any] = faiss.IndexFlat(5 ) lowerCamelCase__ : int = FaissIndex(custom_index=UpperCAmelCase ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def A_ ( self : Any ) -> Optional[int]: import faiss lowerCamelCase__ : Any = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: index.save(tmp_file.name ) lowerCamelCase__ : List[Any] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ : List[str] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Tuple = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(UpperCAmelCase ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: import faiss lowerCamelCase__ : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) lowerCamelCase__ : Optional[int] = 'index.faiss' lowerCamelCase__ : Optional[Any] = F"""mock://{index_name}""" index.save(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Tuple = FaissIndex.load(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Optional[int] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Dict = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(_UpperCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Dict ) -> List[Any]: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : Any = Elasticsearch() lowerCamelCase__ : Tuple = {'acknowledged': True} lowerCamelCase__ : Tuple = ElasticSearchIndex(es_client=UpperCAmelCase ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query lowerCamelCase__ : Optional[int] = 'foo' lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search(UpperCAmelCase ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout lowerCamelCase__ : Any = 'foo' lowerCamelCase__ : List[str] = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Tuple = index.search(UpperCAmelCase , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries lowerCamelCase__ : List[str] = ['foo', 'bar', 'foobar'] lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : str = index.search_batch(UpperCAmelCase ) lowerCamelCase__ : List[str] = [scores[0] for scores in total_scores] lowerCamelCase__ : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase ) # batched queries with timeout lowerCamelCase__ : str = ['foo', 'bar', 'foobar'] lowerCamelCase__ : List[Any] = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search_batch(UpperCAmelCase , request_timeout=30 ) lowerCamelCase__ : Optional[Any] = [scores[0] for scores in total_scores] lowerCamelCase__ : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase )
50
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class lowerCAmelCase ( unittest.TestCase ): def __init__( self : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Any=7 , UpperCAmelCase : Dict=3 , UpperCAmelCase : Optional[int]=10 , UpperCAmelCase : List[Any]=18 , UpperCAmelCase : Dict=30 , UpperCAmelCase : List[str]=400 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : List[str]=[0.5, 0.5, 0.5] , UpperCAmelCase : Optional[int]=[0.5, 0.5, 0.5] , UpperCAmelCase : List[str]=None , ) -> Any: lowerCamelCase__ : Any = size if size is not None else {'shortest_edge': 18} lowerCamelCase__ : List[Any] = crop_size if crop_size is not None else {'height': 18, 'width': 18} lowerCamelCase__ : Tuple = parent lowerCamelCase__ : List[Any] = batch_size lowerCamelCase__ : Dict = num_channels lowerCamelCase__ : Union[str, Any] = num_frames lowerCamelCase__ : Any = image_size lowerCamelCase__ : int = min_resolution lowerCamelCase__ : Tuple = max_resolution lowerCamelCase__ : List[str] = do_resize lowerCamelCase__ : List[str] = size lowerCamelCase__ : Union[str, Any] = do_normalize lowerCamelCase__ : List[str] = image_mean lowerCamelCase__ : Tuple = image_std lowerCamelCase__ : Tuple = crop_size def A_ ( self : Tuple ) -> List[Any]: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = VivitImageProcessor if is_vision_available() else None def A_ ( self : Optional[Any] ) -> Union[str, Any]: lowerCamelCase__ : Optional[int] = VivitImageProcessingTester(self ) @property def A_ ( self : int ) -> Union[str, Any]: return self.image_processor_tester.prepare_image_processor_dict() def A_ ( self : Dict ) -> Any: lowerCamelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'image_std' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_center_crop' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'size' ) ) def A_ ( self : Dict ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) lowerCamelCase__ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def A_ ( self : List[Any] ) -> Optional[int]: # Initialize image_processing lowerCamelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos lowerCamelCase__ : List[Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase ) for video in video_inputs: self.assertIsInstance(UpperCAmelCase , UpperCAmelCase ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input lowerCamelCase__ : str = image_processing(video_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Optional[int] = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A_ ( self : int ) -> Union[str, Any]: # Initialize image_processing lowerCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : List[str] = prepare_video_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , numpify=UpperCAmelCase ) for video in video_inputs: self.assertIsInstance(UpperCAmelCase , UpperCAmelCase ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input lowerCamelCase__ : Any = image_processing(video_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A_ ( self : Dict ) -> Optional[int]: # Initialize image_processing lowerCamelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Union[str, Any] = prepare_video_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , torchify=UpperCAmelCase ) for video in video_inputs: self.assertIsInstance(UpperCAmelCase , UpperCAmelCase ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processing(video_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Tuple = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : str = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCamelCase__ : Tuple = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCamelCase__ : Dict = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCamelCase__ : str = subset[i - 1][j] if arr[i - 1] <= j: lowerCamelCase__ : Dict = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : int = { """configuration_megatron_bert""": ["""MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MegatronBertConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MegatronBertForCausalLM""", """MegatronBertForMaskedLM""", """MegatronBertForMultipleChoice""", """MegatronBertForNextSentencePrediction""", """MegatronBertForPreTraining""", """MegatronBertForQuestionAnswering""", """MegatronBertForSequenceClassification""", """MegatronBertForTokenClassification""", """MegatronBertModel""", """MegatronBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys _UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """M-CLIP""" def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=1024 , UpperCAmelCase : Tuple=768 , **UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[int] = transformerDimSize lowerCamelCase__ : Optional[Any] = imageDimSize super().__init__(**UpperCAmelCase ) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = MCLIPConfig def __init__( self : List[Any] , UpperCAmelCase : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Dict: super().__init__(UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Tuple = XLMRobertaModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict ) -> Tuple: lowerCamelCase__ : Any = self.transformer(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase )[0] lowerCamelCase__ : int = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(UpperCAmelCase ), embs
50
1
import argparse import os import re import packaging.version _UpperCAmelCase : Optional[int] = """examples/""" _UpperCAmelCase : List[str] = { """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } _UpperCAmelCase : int = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } _UpperCAmelCase : List[str] = """README.md""" def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Dict: with open(_UpperCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowerCamelCase__ : Optional[Any] = f.read() lowerCamelCase__ , lowerCamelCase__ : str = REPLACE_PATTERNS[pattern] lowerCamelCase__ : str = replace.replace('VERSION' , _UpperCAmelCase ) lowerCamelCase__ : List[Any] = re_pattern.sub(_UpperCAmelCase , _UpperCAmelCase ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.write(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: for folder, directories, fnames in os.walk(_UpperCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('research_projects' ) if "legacy" in directories: directories.remove('legacy' ) for fname in fnames: if fname.endswith('.py' ): update_version_in_file(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) , _UpperCAmelCase , pattern='examples' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=False ) -> Optional[Any]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if not patch: update_version_in_examples(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : List[str] = '🤗 Transformers currently provides the following architectures' lowerCamelCase__ : str = '1. Want to contribute a new model?' with open(_UpperCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowerCamelCase__ : str = f.readlines() # Find the start of the list. lowerCamelCase__ : Any = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCamelCase__ : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('1.' ): lowerCamelCase__ : str = lines[index].replace( 'https://huggingface.co/docs/transformers/main/model_doc' , 'https://huggingface.co/docs/transformers/model_doc' , ) index += 1 with open(_UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> Optional[int]: with open(REPLACE_FILES['init'] , 'r' ) as f: lowerCamelCase__ : Any = f.read() lowerCamelCase__ : Union[str, Any] = REPLACE_PATTERNS['init'][0].search(_UpperCAmelCase ).groups()[0] return packaging.version.parse(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=False ) -> Optional[Any]: lowerCamelCase__ : Tuple = get_version() if patch and default_version.is_devrelease: raise ValueError('Can\'t create a patch version from the dev branch, checkout a released version!' ) if default_version.is_devrelease: lowerCamelCase__ : List[str] = default_version.base_version elif patch: lowerCamelCase__ : Union[str, Any] = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowerCamelCase__ : Union[str, Any] = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowerCamelCase__ : Optional[int] = input(F"""Which version are you releasing? [{default_version}]""" ) if len(_UpperCAmelCase ) == 0: lowerCamelCase__ : str = default_version print(F"""Updating version to {version}.""" ) global_version_update(_UpperCAmelCase , patch=_UpperCAmelCase ) if not patch: print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() def SCREAMING_SNAKE_CASE ( ) -> Tuple: lowerCamelCase__ : Tuple = get_version() lowerCamelCase__ : List[Any] = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowerCamelCase__ : Union[str, Any] = current_version.base_version # Check with the user we got that right. lowerCamelCase__ : Any = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(_UpperCAmelCase ) == 0: lowerCamelCase__ : Dict = dev_version print(F"""Updating version to {version}.""" ) global_version_update(_UpperCAmelCase ) print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() if __name__ == "__main__": _UpperCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") _UpperCAmelCase : int = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
50
from itertools import count def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Optional[Any] = [1] * min_block_length for n in count(_UpperCAmelCase ): fill_count_functions.append(1 ) for block_length in range(_UpperCAmelCase , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
50
1
import baseaa def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bytes: return baseaa.baaencode(string.encode('utf-8' ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: return baseaa.baadecode(_UpperCAmelCase ).decode('utf-8' ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = """Hello World!""" _UpperCAmelCase : int = baseaa_encode(test) print(encoded) _UpperCAmelCase : Optional[int] = baseaa_decode(encoded) print(decoded)
50
from __future__ import annotations from typing import Any def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: create_state_space_tree(_UpperCAmelCase , [] , 0 ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> None: if index == len(_UpperCAmelCase ): print(_UpperCAmelCase ) return create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _UpperCAmelCase : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["""A""", """B""", """C"""]) generate_all_subsequences(seq)
50
1
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _UpperCAmelCase : Union[str, Any] = False try: _UpperCAmelCase : Optional[Any] = _is_package_available("""google.colab""") except ModuleNotFoundError: pass @input.register class lowerCAmelCase : def __init__( self : List[Any] , UpperCAmelCase : str = None , UpperCAmelCase : list = [] ) -> Optional[int]: lowerCamelCase__ : Optional[int] = 0 lowerCamelCase__ : Dict = choices lowerCamelCase__ : Union[str, Any] = prompt if sys.platform == "win32": lowerCamelCase__ : List[str] = '*' else: lowerCamelCase__ : Optional[Any] = '➔ ' def A_ ( self : Optional[int] , UpperCAmelCase : Optional[int] , UpperCAmelCase : str = "" ) -> Tuple: if sys.platform != "win32": writeColor(self.choices[index] , 32 , UpperCAmelCase ) else: forceWrite(self.choices[index] , UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : int ) -> Any: if index == self.position: forceWrite(F""" {self.arrow_char} """ ) self.write_choice(UpperCAmelCase ) else: forceWrite(F""" {self.choices[index]}""" ) reset_cursor() def A_ ( self : Dict , UpperCAmelCase : Direction , UpperCAmelCase : int = 1 ) -> Any: lowerCamelCase__ : int = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(UpperCAmelCase ) move_cursor(UpperCAmelCase , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP['up'] ) def A_ ( self : Optional[Any] ) -> List[str]: self.move_direction(Direction.UP ) @input.mark(KEYMAP['down'] ) def A_ ( self : List[str] ) -> int: self.move_direction(Direction.DOWN ) @input.mark(KEYMAP['newline'] ) def A_ ( self : str ) -> List[Any]: move_cursor(len(self.choices ) - self.position , 'DOWN' ) return self.position @input.mark(KEYMAP['interrupt'] ) def A_ ( self : Tuple ) -> Any: move_cursor(len(self.choices ) - self.position , 'DOWN' ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(UpperCAmelCase )] for number in range(10 )] ) def A_ ( self : Optional[Any] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = int(chr(self.current_selection ) ) lowerCamelCase__ : Union[str, Any] = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , UpperCAmelCase ) else: return else: return def A_ ( self : Optional[Any] , UpperCAmelCase : int = 0 ) -> str: if self.prompt: linebreak() forceWrite(self.prompt , '\n' ) if in_colab: forceWrite('Please input a choice index (starting from 0), and press enter' , '\n' ) else: forceWrite('Please select a choice using the arrow or number keys, and selecting with enter' , '\n' ) lowerCamelCase__ : int = default_choice for i in range(len(self.choices ) ): self.print_choice(UpperCAmelCase ) forceWrite('\n' ) move_cursor(len(self.choices ) - self.position , 'UP' ) with cursor.hide(): while True: if in_colab: try: lowerCamelCase__ : Any = int(builtins.input() ) except ValueError: lowerCamelCase__ : Optional[int] = default_choice else: lowerCamelCase__ : Union[str, Any] = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , 'UP' ) clear_line() self.write_choice(UpperCAmelCase , '\n' ) return choice
50
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off _UpperCAmelCase : Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = NllbTokenizer UpperCAmelCase__ = [] UpperCAmelCase__ = [] def __init__( self : Tuple , UpperCAmelCase : int=None , UpperCAmelCase : Any=None , UpperCAmelCase : str="<s>" , UpperCAmelCase : Optional[Any]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Tuple="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[str]="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=False , **UpperCAmelCase : Optional[int] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token lowerCamelCase__ : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase , tokenizer_file=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , src_lang=UpperCAmelCase , tgt_lang=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , legacy_behaviour=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = vocab_file lowerCamelCase__ : Dict = False if not self.vocab_file else True lowerCamelCase__ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCamelCase__ : str = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : int = src_lang if src_lang is not None else 'eng_Latn' lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A_ ( self : int ) -> str: return self._src_lang @src_lang.setter def A_ ( self : List[Any] , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] , UpperCAmelCase : Optional[str] , **UpperCAmelCase : List[str] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCamelCase__ : Optional[int] = src_lang lowerCamelCase__ : Optional[int] = self(UpperCAmelCase , add_special_tokens=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.convert_tokens_to_ids(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tgt_lang_id return inputs def A_ ( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : str = "eng_Latn" , UpperCAmelCase : Optional[List[str]] = None , UpperCAmelCase : str = "fra_Latn" , **UpperCAmelCase : Dict , ) -> BatchEncoding: lowerCamelCase__ : Any = src_lang lowerCamelCase__ : int = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Optional[int]: return self.set_src_lang_special_tokens(self.src_lang ) def A_ ( self : Any ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A_ ( self : str , UpperCAmelCase : Optional[Any] ) -> None: lowerCamelCase__ : int = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : int = [self.cur_lang_code] lowerCamelCase__ : Tuple = [self.eos_token_id] lowerCamelCase__ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : int , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Union[str, Any] = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : Dict = [] lowerCamelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : Any = [self.cur_lang_code] lowerCamelCase__ : Optional[Any] = [self.eos_token_id] lowerCamelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCamelCase__ : int = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
1
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase : def __init__( self : int , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any]=13 , UpperCAmelCase : Optional[int]=7 , UpperCAmelCase : int=True , UpperCAmelCase : Any=True , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : List[Any]=True , UpperCAmelCase : int=99 , UpperCAmelCase : Tuple=24 , UpperCAmelCase : Any=2 , UpperCAmelCase : Tuple=6 , UpperCAmelCase : int=37 , UpperCAmelCase : Optional[Any]="gelu" , UpperCAmelCase : Optional[Any]=0.1 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : str=512 , UpperCAmelCase : List[str]=16 , UpperCAmelCase : int=2 , UpperCAmelCase : List[str]=0.0_2 , UpperCAmelCase : int=3 , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : str=1000 , ) -> int: lowerCamelCase__ : List[str] = parent lowerCamelCase__ : Union[str, Any] = batch_size lowerCamelCase__ : Tuple = seq_length lowerCamelCase__ : Any = is_training lowerCamelCase__ : Optional[int] = use_input_mask lowerCamelCase__ : Any = use_token_type_ids lowerCamelCase__ : int = use_labels lowerCamelCase__ : str = vocab_size lowerCamelCase__ : int = hidden_size lowerCamelCase__ : Dict = num_hidden_layers lowerCamelCase__ : List[Any] = num_attention_heads lowerCamelCase__ : str = intermediate_size lowerCamelCase__ : int = hidden_act lowerCamelCase__ : str = hidden_dropout_prob lowerCamelCase__ : List[Any] = attention_probs_dropout_prob lowerCamelCase__ : Union[str, Any] = max_position_embeddings lowerCamelCase__ : Optional[Any] = type_vocab_size lowerCamelCase__ : int = type_sequence_label_size lowerCamelCase__ : str = initializer_range lowerCamelCase__ : List[Any] = num_labels lowerCamelCase__ : List[str] = scope lowerCamelCase__ : str = range_bbox def A_ ( self : Tuple ) -> Tuple: lowerCamelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCamelCase__ : Union[str, Any] = bbox[i, j, 3] lowerCamelCase__ : List[str] = bbox[i, j, 1] lowerCamelCase__ : Any = t if bbox[i, j, 2] < bbox[i, j, 0]: lowerCamelCase__ : Optional[int] = bbox[i, j, 2] lowerCamelCase__ : str = bbox[i, j, 0] lowerCamelCase__ : str = t lowerCamelCase__ : Any = None if self.use_input_mask: lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowerCamelCase__ : List[Any] = None if self.use_token_type_ids: lowerCamelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : List[str] = None if self.use_labels: lowerCamelCase__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase__ : Tuple = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def A_ ( self : List[str] ) -> Tuple: return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def A_ ( self : Optional[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Any , UpperCAmelCase : Any , UpperCAmelCase : List[str] , UpperCAmelCase : Any , ) -> List[str]: lowerCamelCase__ : str = LiltModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : str = model(UpperCAmelCase , bbox=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , bbox=UpperCAmelCase , token_type_ids=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase , bbox=UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A_ ( self : Tuple , UpperCAmelCase : str , UpperCAmelCase : List[Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple , ) -> int: lowerCamelCase__ : Tuple = self.num_labels lowerCamelCase__ : List[str] = LiltForTokenClassification(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Dict = model( UpperCAmelCase , bbox=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A_ ( self : List[str] , UpperCAmelCase : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , UpperCAmelCase : str , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] , ) -> Any: lowerCamelCase__ : Optional[int] = LiltForQuestionAnswering(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( UpperCAmelCase , bbox=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , start_positions=UpperCAmelCase , end_positions=UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A_ ( self : Tuple ) -> Tuple: lowerCamelCase__ : Tuple = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Optional[int] = config_and_inputs lowerCamelCase__ : Any = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) UpperCAmelCase__ = ( { """feature-extraction""": LiltModel, """question-answering""": LiltForQuestionAnswering, """text-classification""": LiltForSequenceClassification, """token-classification""": LiltForTokenClassification, """zero-shot""": LiltForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : Dict , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Any , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> Tuple: return True def A_ ( self : Union[str, Any] ) -> int: lowerCamelCase__ : int = LiltModelTester(self ) lowerCamelCase__ : Tuple = ConfigTester(self , config_class=UpperCAmelCase , hidden_size=37 ) def A_ ( self : int ) -> Tuple: self.config_tester.run_common_tests() def A_ ( self : int ) -> Any: lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def A_ ( self : Dict ) -> List[Any]: lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase__ : Union[str, Any] = type self.model_tester.create_and_check_model(*UpperCAmelCase ) def A_ ( self : Optional[int] ) -> Tuple: lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Tuple: lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase ) @slow def A_ ( self : List[str] ) -> Tuple: for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Optional[int] = LiltModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @require_torch @slow class lowerCAmelCase ( unittest.TestCase ): def A_ ( self : str ) -> Dict: lowerCamelCase__ : List[str] = LiltModel.from_pretrained('SCUT-DLVCLab/lilt-roberta-en-base' ).to(UpperCAmelCase ) lowerCamelCase__ : List[Any] = torch.tensor([[1, 2]] , device=UpperCAmelCase ) lowerCamelCase__ : List[Any] = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=UpperCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(input_ids=UpperCAmelCase , bbox=UpperCAmelCase ) lowerCamelCase__ : List[str] = torch.Size([1, 2, 768] ) lowerCamelCase__ : Any = torch.tensor( [[-0.0_6_5_3, 0.0_9_5_0, -0.0_0_6_1], [-0.0_5_4_5, 0.0_9_2_6, -0.0_3_2_4]] , device=UpperCAmelCase , ) self.assertTrue(outputs.last_hidden_state.shape , UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , UpperCAmelCase , atol=1e-3 ) )
50
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ , lowerCamelCase__ : List[str] = emb.weight.shape lowerCamelCase__ : Tuple = nn.Linear(_UpperCAmelCase , _UpperCAmelCase , bias=_UpperCAmelCase ) lowerCamelCase__ : Dict = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location='cpu' ) lowerCamelCase__ : List[str] = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCamelCase__ : Optional[int] = mam_aaa['model'] remove_ignore_keys_(_UpperCAmelCase ) lowerCamelCase__ : str = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCamelCase__ : Union[str, Any] = MaMaaaConfig( vocab_size=_UpperCAmelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCamelCase__ : Optional[Any] = state_dict['decoder.embed_tokens.weight'] lowerCamelCase__ : Union[str, Any] = MaMaaaForConditionalGeneration(_UpperCAmelCase ) model.model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) lowerCamelCase__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase : str = parser.parse_args() _UpperCAmelCase : Optional[Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
50
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} _UpperCAmelCase : Dict = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } _UpperCAmelCase : List[Any] = { """albert-base-v1""": 5_12, """albert-large-v1""": 5_12, """albert-xlarge-v1""": 5_12, """albert-xxlarge-v1""": 5_12, """albert-base-v2""": 5_12, """albert-large-v2""": 5_12, """albert-xlarge-v2""": 5_12, """albert-xxlarge-v2""": 5_12, } _UpperCAmelCase : Optional[int] = """▁""" class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[Any] , UpperCAmelCase : List[str] , UpperCAmelCase : str=True , UpperCAmelCase : List[str]=True , UpperCAmelCase : str=False , UpperCAmelCase : str="[CLS]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Union[str, Any]="<unk>" , UpperCAmelCase : int="[SEP]" , UpperCAmelCase : str="<pad>" , UpperCAmelCase : Dict="[CLS]" , UpperCAmelCase : str="[MASK]" , UpperCAmelCase : Optional[Dict[str, Any]] = None , **UpperCAmelCase : Optional[int] , ) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCamelCase__ : int = ( AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase , normalized=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token ) lowerCamelCase__ : Dict = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase , remove_space=UpperCAmelCase , keep_accents=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase , ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : List[str] = remove_space lowerCamelCase__ : Any = keep_accents lowerCamelCase__ : Dict = vocab_file lowerCamelCase__ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase ) @property def A_ ( self : Any ) -> Optional[int]: return len(self.sp_model ) def A_ ( self : List[str] ) -> Optional[Any]: lowerCamelCase__ : Dict = {self.convert_ids_to_tokens(UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ) -> int: lowerCamelCase__ : List[Any] = self.__dict__.copy() lowerCamelCase__ : List[Any] = None return state def __setstate__( self : List[str] , UpperCAmelCase : Any ) -> Optional[int]: lowerCamelCase__ : List[Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCamelCase__ : Dict = {} lowerCamelCase__ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A_ ( self : Dict , UpperCAmelCase : Union[str, Any] ) -> Tuple: if self.remove_space: lowerCamelCase__ : int = ' '.join(inputs.strip().split() ) else: lowerCamelCase__ : int = inputs lowerCamelCase__ : List[Any] = outputs.replace('``' , '"' ).replace('\'\'' , '"' ) if not self.keep_accents: lowerCamelCase__ : List[Any] = unicodedata.normalize('NFKD' , UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = ''.join([c for c in outputs if not unicodedata.combining(UpperCAmelCase )] ) if self.do_lower_case: lowerCamelCase__ : Any = outputs.lower() return outputs def A_ ( self : List[Any] , UpperCAmelCase : str ) -> List[str]: lowerCamelCase__ : Optional[int] = self.preprocess_text(UpperCAmelCase ) lowerCamelCase__ : Dict = self.sp_model.encode(UpperCAmelCase , out_type=UpperCAmelCase ) lowerCamelCase__ : List[Any] = [] for piece in pieces: if len(UpperCAmelCase ) > 1 and piece[-1] == str(',' ) and piece[-2].isdigit(): lowerCamelCase__ : str = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase , '' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase__ : List[str] = cur_pieces[1:] else: lowerCamelCase__ : List[Any] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase ) else: new_pieces.append(UpperCAmelCase ) return new_pieces def A_ ( self : Dict , UpperCAmelCase : Optional[int] ) -> List[Any]: return self.sp_model.PieceToId(UpperCAmelCase ) def A_ ( self : Optional[Any] , UpperCAmelCase : Optional[Any] ) -> Any: return self.sp_model.IdToPiece(UpperCAmelCase ) def A_ ( self : str , UpperCAmelCase : Dict ) -> Optional[int]: lowerCamelCase__ : str = [] lowerCamelCase__ : Tuple = '' lowerCamelCase__ : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(UpperCAmelCase ) + token lowerCamelCase__ : List[Any] = True lowerCamelCase__ : str = [] else: current_sub_tokens.append(UpperCAmelCase ) lowerCamelCase__ : List[str] = False out_string += self.sp_model.decode(UpperCAmelCase ) return out_string.strip() def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Optional[Any] = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A_ ( self : Optional[int] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None , UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase , token_ids_a=UpperCAmelCase , already_has_special_tokens=UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(UpperCAmelCase )) + [1] + ([0] * len(UpperCAmelCase )) + [1] return [1] + ([0] * len(UpperCAmelCase )) + [1] def A_ ( self : str , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Optional[int] = [self.sep_token_id] lowerCamelCase__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : Dict = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase , 'wb' ) as fi: lowerCamelCase__ : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase ) return (out_vocab_file,)
50
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
1
from collections.abc import Callable def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> float: lowerCamelCase__ : float = a lowerCamelCase__ : float = b if function(_UpperCAmelCase ) == 0: # one of the a or b is a root for the function return a elif function(_UpperCAmelCase ) == 0: return b elif ( function(_UpperCAmelCase ) * function(_UpperCAmelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError('could not find root in given interval.' ) else: lowerCamelCase__ : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(_UpperCAmelCase ) == 0: return mid elif function(_UpperCAmelCase ) * function(_UpperCAmelCase ) < 0: lowerCamelCase__ : Optional[int] = mid else: lowerCamelCase__ : Optional[Any] = mid lowerCamelCase__ : Any = start + (end - start) / 2.0 return mid def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> float: return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 10_00)) import doctest doctest.testmod()
50
import flax.linen as nn import jax import jax.numpy as jnp class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , UpperCAmelCase : Dict ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = hidden_states.shape lowerCamelCase__ : Union[str, Any] = jax.image.resize( UpperCAmelCase , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : str , UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = None UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels lowerCamelCase__ : Tuple = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : int = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Union[str, Any] = nn.Dense(UpperCAmelCase , dtype=self.dtype ) lowerCamelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : List[Any] = nn.Dropout(self.dropout_prob ) lowerCamelCase__ : Tuple = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Optional[Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCamelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCamelCase__ : Dict = nn.Conv( UpperCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=True ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = hidden_states lowerCamelCase__ : List[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[Any] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Any = self.conva(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.time_emb_proj(nn.swish(UpperCAmelCase ) ) lowerCamelCase__ : List[str] = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase , 1 ) , 1 ) lowerCamelCase__ : List[str] = hidden_states + temb lowerCamelCase__ : Optional[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[str] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.dropout(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : str = self.conva(UpperCAmelCase ) if self.conv_shortcut is not None: lowerCamelCase__ : Dict = self.conv_shortcut(UpperCAmelCase ) return hidden_states + residual
50
1
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : int = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/vocab.json""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/vocab.json""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/vocab.json""", """roberta-base-openai-detector""": """https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json""", """roberta-large-openai-detector""": ( """https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json""" ), }, """merges_file""": { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/merges.txt""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/merges.txt""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/merges.txt""", """roberta-base-openai-detector""": """https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt""", """roberta-large-openai-detector""": ( """https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt""" ), }, """tokenizer_file""": { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/tokenizer.json""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/tokenizer.json""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json""", """roberta-base-openai-detector""": ( """https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json""" ), """roberta-large-openai-detector""": ( """https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : Optional[int] = { """roberta-base""": 5_12, """roberta-large""": 5_12, """roberta-large-mnli""": 5_12, """distilroberta-base""": 5_12, """roberta-base-openai-detector""": 5_12, """roberta-large-openai-detector""": 5_12, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = RobertaTokenizer def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[Any]="replace" , UpperCAmelCase : Any="<s>" , UpperCAmelCase : Union[str, Any]="</s>" , UpperCAmelCase : int="</s>" , UpperCAmelCase : Optional[Any]="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[Any]="<pad>" , UpperCAmelCase : int="<mask>" , UpperCAmelCase : int=False , UpperCAmelCase : str=True , **UpperCAmelCase : Tuple , ) -> List[str]: super().__init__( UpperCAmelCase , UpperCAmelCase , tokenizer_file=UpperCAmelCase , errors=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , add_prefix_space=UpperCAmelCase , trim_offsets=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , UpperCAmelCase ) != add_prefix_space: lowerCamelCase__ : int = getattr(UpperCAmelCase , pre_tok_state.pop('type' ) ) lowerCamelCase__ : int = add_prefix_space lowerCamelCase__ : Tuple = pre_tok_class(**UpperCAmelCase ) lowerCamelCase__ : List[str] = add_prefix_space lowerCamelCase__ : str = 'post_processor' lowerCamelCase__ : Union[str, Any] = getattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) if tokenizer_component_instance: lowerCamelCase__ : str = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : int = tuple(state['sep'] ) if "cls" in state: lowerCamelCase__ : str = tuple(state['cls'] ) lowerCamelCase__ : Union[str, Any] = False if state.get('add_prefix_space' , UpperCAmelCase ) != add_prefix_space: lowerCamelCase__ : int = add_prefix_space lowerCamelCase__ : int = True if state.get('trim_offsets' , UpperCAmelCase ) != trim_offsets: lowerCamelCase__ : List[str] = trim_offsets lowerCamelCase__ : Union[str, Any] = True if changes_to_apply: lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , state.pop('type' ) ) lowerCamelCase__ : str = component_class(**UpperCAmelCase ) setattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) @property def A_ ( self : Optional[int] ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def A_ ( self : Optional[int] , UpperCAmelCase : List[str] ) -> str: lowerCamelCase__ : List[str] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else value lowerCamelCase__ : Union[str, Any] = value def A_ ( self : int , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Any ) -> BatchEncoding: lowerCamelCase__ : Optional[Any] = kwargs.get('is_split_into_words' , UpperCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Optional[int] , *UpperCAmelCase : int , **UpperCAmelCase : Optional[Any] ) -> BatchEncoding: lowerCamelCase__ : Optional[Any] = kwargs.get('is_split_into_words' , UpperCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Any , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Optional[Any]=None ) -> Dict: lowerCamelCase__ : Any = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A_ ( self : Any , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[str] = [self.sep_token_id] lowerCamelCase__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Optional[Any] = set() # edges = list of graph's edges lowerCamelCase__ : List[str] = get_edges(_UpperCAmelCase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCamelCase__ , lowerCamelCase__ : str = edges.pop() chosen_vertices.add(_UpperCAmelCase ) chosen_vertices.add(_UpperCAmelCase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_UpperCAmelCase ) return chosen_vertices def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Union[str, Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
50
1
from math import factorial _UpperCAmelCase : dict[str, int] = {str(digit): factorial(digit) for digit in range(10)} def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('Parameter number must be int' ) if number < 0: raise ValueError('Parameter number must be greater than or equal to 0' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 60 , _UpperCAmelCase = 100_0000 ) -> int: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ) or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('Parameters chain_length and number_limit must be int' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( 'Parameters chain_length and number_limit must be greater than 0' ) # the counter for the chains with the exact desired length lowerCamelCase__ : Optional[Any] = 0 # the cached sizes of the previous chains lowerCamelCase__ : dict[int, int] = {} for start_chain_element in range(1 , _UpperCAmelCase ): # The temporary set will contain the elements of the chain lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. lowerCamelCase__ : Tuple = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(_UpperCAmelCase ) chain_set_length += 1 lowerCamelCase__ : Union[str, Any] = digit_factorial_sum(_UpperCAmelCase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] lowerCamelCase__ : int = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
50
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
1
import os _UpperCAmelCase : int = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 1_00, """D""": 5_00, """M""": 10_00} def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Optional[Any] = 0 while index < len(_UpperCAmelCase ) - 1: lowerCamelCase__ : Optional[int] = SYMBOLS[numerals[index]] lowerCamelCase__ : Optional[Any] = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : int = '' lowerCamelCase__ : Any = num // 1000 numerals += m_count * "M" num %= 1000 lowerCamelCase__ : str = num // 100 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 100 lowerCamelCase__ : List[str] = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = "/p089_roman.txt" ) -> int: lowerCamelCase__ : Tuple = 0 with open(os.path.dirname(_UpperCAmelCase ) + roman_numerals_filename ) as filea: lowerCamelCase__ : List[str] = filea.readlines() for line in lines: lowerCamelCase__ : Optional[int] = line.strip() lowerCamelCase__ : Union[str, Any] = parse_roman_numerals(_UpperCAmelCase ) lowerCamelCase__ : List[Any] = generate_roman_numerals(_UpperCAmelCase ) savings += len(_UpperCAmelCase ) - len(_UpperCAmelCase ) return savings if __name__ == "__main__": print(F"""{solution() = }""")
50
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
1
import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase="shi-labs/oneformer_demo" ) -> List[Any]: with open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) as f: lowerCamelCase__ : str = json.load(_UpperCAmelCase ) lowerCamelCase__ : Tuple = {} lowerCamelCase__ : List[Any] = [] lowerCamelCase__ : str = [] for key, info in class_info.items(): lowerCamelCase__ : Union[str, Any] = info['name'] class_names.append(info['name'] ) if info["isthing"]: thing_ids.append(int(_UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = thing_ids lowerCamelCase__ : Union[str, Any] = class_names return metadata class lowerCAmelCase ( unittest.TestCase ): def __init__( self : List[str] , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int]=7 , UpperCAmelCase : Union[str, Any]=3 , UpperCAmelCase : int=30 , UpperCAmelCase : List[str]=400 , UpperCAmelCase : List[Any]=None , UpperCAmelCase : Any=True , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Dict=[0.5, 0.5, 0.5] , UpperCAmelCase : List[str]=[0.5, 0.5, 0.5] , UpperCAmelCase : Optional[Any]=10 , UpperCAmelCase : Tuple=False , UpperCAmelCase : Optional[int]=255 , UpperCAmelCase : Any="shi-labs/oneformer_demo" , UpperCAmelCase : Any="ade20k_panoptic.json" , UpperCAmelCase : List[Any]=10 , ) -> Union[str, Any]: lowerCamelCase__ : Tuple = parent lowerCamelCase__ : Tuple = batch_size lowerCamelCase__ : str = num_channels lowerCamelCase__ : Union[str, Any] = min_resolution lowerCamelCase__ : int = max_resolution lowerCamelCase__ : Dict = do_resize lowerCamelCase__ : Optional[int] = {'shortest_edge': 32, 'longest_edge': 1333} if size is None else size lowerCamelCase__ : Dict = do_normalize lowerCamelCase__ : Tuple = image_mean lowerCamelCase__ : List[str] = image_std lowerCamelCase__ : Any = class_info_file lowerCamelCase__ : Any = prepare_metadata(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = num_text lowerCamelCase__ : List[str] = repo_path # for the post_process_functions lowerCamelCase__ : Any = 2 lowerCamelCase__ : str = 10 lowerCamelCase__ : str = 10 lowerCamelCase__ : Any = 3 lowerCamelCase__ : Union[str, Any] = 4 lowerCamelCase__ : Any = num_labels lowerCamelCase__ : str = do_reduce_labels lowerCamelCase__ : str = ignore_index def A_ ( self : Union[str, Any] ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def A_ ( self : Optional[int] , UpperCAmelCase : List[str] , UpperCAmelCase : List[Any]=False ) -> int: if not batched: lowerCamelCase__ : List[str] = image_inputs[0] if isinstance(UpperCAmelCase , Image.Image ): lowerCamelCase__ , lowerCamelCase__ : Tuple = image.size else: lowerCamelCase__ , lowerCamelCase__ : Dict = image.shape[1], image.shape[2] if w < h: lowerCamelCase__ : Dict = int(self.size['shortest_edge'] * h / w ) lowerCamelCase__ : List[Any] = self.size['shortest_edge'] elif w > h: lowerCamelCase__ : Optional[Any] = self.size['shortest_edge'] lowerCamelCase__ : str = int(self.size['shortest_edge'] * w / h ) else: lowerCamelCase__ : str = self.size['shortest_edge'] lowerCamelCase__ : Union[str, Any] = self.size['shortest_edge'] else: lowerCamelCase__ : Any = [] for image in image_inputs: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase__ : Optional[Any] = max(UpperCAmelCase , key=lambda UpperCAmelCase : item[0] )[0] lowerCamelCase__ : str = max(UpperCAmelCase , key=lambda UpperCAmelCase : item[1] )[1] return expected_height, expected_width def A_ ( self : Tuple ) -> Tuple: return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string UpperCAmelCase__ = image_processing_class def A_ ( self : Any ) -> int: lowerCamelCase__ : Union[str, Any] = OneFormerImageProcessorTester(self ) @property def A_ ( self : str ) -> int: return self.image_processing_tester.prepare_image_processor_dict() def A_ ( self : int ) -> Any: lowerCamelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'image_std' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'size' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'ignore_index' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'class_info_file' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'num_text' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'repo_path' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'metadata' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_reduce_labels' ) ) def A_ ( self : str ) -> List[Any]: pass def A_ ( self : Tuple ) -> Union[str, Any]: # Initialize image_processor lowerCamelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : int = prepare_image_inputs(self.image_processing_tester , equal_resolution=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , Image.Image ) # Test not batched input lowerCamelCase__ : List[str] = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values lowerCamelCase__ , lowerCamelCase__ : str = self.image_processing_tester.get_expected_values(UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ , lowerCamelCase__ : Tuple = self.image_processing_tester.get_expected_values(UpperCAmelCase , batched=UpperCAmelCase ) lowerCamelCase__ : List[str] = image_processor( UpperCAmelCase , ['semantic'] * len(UpperCAmelCase ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A_ ( self : Tuple ) -> str: # Initialize image_processor lowerCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Union[str, Any] = prepare_image_inputs(self.image_processing_tester , equal_resolution=UpperCAmelCase , numpify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , np.ndarray ) # Test not batched input lowerCamelCase__ : List[str] = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values lowerCamelCase__ , lowerCamelCase__ : List[str] = self.image_processing_tester.get_expected_values(UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ , lowerCamelCase__ : List[str] = self.image_processing_tester.get_expected_values(UpperCAmelCase , batched=UpperCAmelCase ) lowerCamelCase__ : str = image_processor( UpperCAmelCase , ['semantic'] * len(UpperCAmelCase ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A_ ( self : Optional[int] ) -> Union[str, Any]: # Initialize image_processor lowerCamelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Union[str, Any] = prepare_image_inputs(self.image_processing_tester , equal_resolution=UpperCAmelCase , torchify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , torch.Tensor ) # Test not batched input lowerCamelCase__ : Union[str, Any] = image_processor(image_inputs[0] , ['semantic'] , return_tensors='pt' ).pixel_values lowerCamelCase__ , lowerCamelCase__ : str = self.image_processing_tester.get_expected_values(UpperCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase__ , lowerCamelCase__ : int = self.image_processing_tester.get_expected_values(UpperCAmelCase , batched=UpperCAmelCase ) lowerCamelCase__ : int = image_processor( UpperCAmelCase , ['semantic'] * len(UpperCAmelCase ) , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A_ ( self : int , UpperCAmelCase : List[str]=False , UpperCAmelCase : List[Any]=False , UpperCAmelCase : Union[str, Any]="np" ) -> str: lowerCamelCase__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # prepare image and target lowerCamelCase__ : Dict = self.image_processing_tester.num_labels lowerCamelCase__ : List[str] = None lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : str = prepare_image_inputs(self.image_processing_tester , equal_resolution=UpperCAmelCase ) if with_segmentation_maps: lowerCamelCase__ : Tuple = num_labels if is_instance_map: lowerCamelCase__ : Dict = list(range(UpperCAmelCase ) ) * 2 lowerCamelCase__ : Optional[int] = dict(enumerate(UpperCAmelCase ) ) lowerCamelCase__ : int = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": lowerCamelCase__ : Optional[int] = [Image.fromarray(UpperCAmelCase ) for annotation in annotations] lowerCamelCase__ : List[str] = image_processor( UpperCAmelCase , ['semantic'] * len(UpperCAmelCase ) , UpperCAmelCase , return_tensors='pt' , instance_id_to_semantic_id=UpperCAmelCase , pad_and_return_pixel_mask=UpperCAmelCase , ) return inputs def A_ ( self : str ) -> Any: pass def A_ ( self : Tuple ) -> List[Any]: def common(UpperCAmelCase : Union[str, Any]=False , UpperCAmelCase : Optional[Any]=None ): lowerCamelCase__ : Any = self.comm_get_image_processor_inputs( with_segmentation_maps=UpperCAmelCase , is_instance_map=UpperCAmelCase , segmentation_type=UpperCAmelCase ) lowerCamelCase__ : Tuple = inputs['mask_labels'] lowerCamelCase__ : Union[str, Any] = inputs['class_labels'] lowerCamelCase__ : Optional[Any] = inputs['pixel_values'] lowerCamelCase__ : List[Any] = inputs['text_inputs'] # check the batch_size for mask_label, class_label, text_input in zip(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(UpperCAmelCase ) , self.image_processing_tester.num_text ) common() common(is_instance_map=UpperCAmelCase ) common(is_instance_map=UpperCAmelCase , segmentation_type='pil' ) common(is_instance_map=UpperCAmelCase , segmentation_type='pil' ) def A_ ( self : Optional[int] ) -> Any: lowerCamelCase__ : Dict = np.zeros((20, 50) ) lowerCamelCase__ : List[Any] = 1 lowerCamelCase__ : Dict = 1 lowerCamelCase__ : Optional[int] = 1 lowerCamelCase__ : Union[str, Any] = binary_mask_to_rle(UpperCAmelCase ) self.assertEqual(len(UpperCAmelCase ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def A_ ( self : Union[str, Any] ) -> str: lowerCamelCase__ : str = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) lowerCamelCase__ : Tuple = self.image_processing_tester.get_fake_oneformer_outputs() lowerCamelCase__ : Any = fature_extractor.post_process_semantic_segmentation(UpperCAmelCase ) self.assertEqual(len(UpperCAmelCase ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) lowerCamelCase__ : Any = [(1, 4) for i in range(self.image_processing_tester.batch_size )] lowerCamelCase__ : Dict = fature_extractor.post_process_semantic_segmentation(UpperCAmelCase , target_sizes=UpperCAmelCase ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def A_ ( self : List[str] ) -> List[str]: lowerCamelCase__ : Tuple = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) lowerCamelCase__ : Union[str, Any] = self.image_processing_tester.get_fake_oneformer_outputs() lowerCamelCase__ : str = image_processor.post_process_instance_segmentation(UpperCAmelCase , threshold=0 ) self.assertTrue(len(UpperCAmelCase ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('segmentation' in el ) self.assertTrue('segments_info' in el ) self.assertEqual(type(el['segments_info'] ) , UpperCAmelCase ) self.assertEqual( el['segmentation'].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def A_ ( self : Any ) -> Union[str, Any]: lowerCamelCase__ : int = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file='ade20k_panoptic.json' , num_text=self.image_processing_tester.num_text , repo_path='shi-labs/oneformer_demo' , ) lowerCamelCase__ : int = self.image_processing_tester.get_fake_oneformer_outputs() lowerCamelCase__ : Tuple = image_processor.post_process_panoptic_segmentation(UpperCAmelCase , threshold=0 ) self.assertTrue(len(UpperCAmelCase ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue('segmentation' in el ) self.assertTrue('segments_info' in el ) self.assertEqual(type(el['segments_info'] ) , UpperCAmelCase ) self.assertEqual( el['segmentation'].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
50
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : str=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : str=9 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=True , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : List[str]=5 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Union[str, Any]=37 , UpperCAmelCase : int=8 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=0.0_0_2 , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=0 , UpperCAmelCase : Union[str, Any]=0 , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , ) -> Union[str, Any]: lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = encoder_seq_length lowerCamelCase__ : int = decoder_seq_length # For common tests lowerCamelCase__ : List[str] = self.decoder_seq_length lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : List[Any] = use_attention_mask lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = d_ff lowerCamelCase__ : Optional[Any] = relative_attention_num_buckets lowerCamelCase__ : Any = dropout_rate lowerCamelCase__ : Any = initializer_factor lowerCamelCase__ : Union[str, Any] = eos_token_id lowerCamelCase__ : List[str] = pad_token_id lowerCamelCase__ : List[str] = decoder_start_token_id lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = decoder_layers def A_ ( self : List[Any] ) -> int: return TaConfig.from_pretrained('google/umt5-base' ) def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=None , ) -> List[str]: if attention_mask is None: lowerCamelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCamelCase__ : Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCamelCase__ : int = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase ) if decoder_head_mask is None: lowerCamelCase__ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A_ ( self : str ) -> List[str]: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCamelCase__ : List[str] = input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Union[str, Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Dict = self.get_config() lowerCamelCase__ : Tuple = config.num_attention_heads lowerCamelCase__ : Any = self.prepare_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, input_dict def A_ ( self : Tuple ) -> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Dict = self.prepare_config_and_inputs() return config, inputs_dict def A_ ( self : Optional[int] ) -> List[str]: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Union[str, Any] ) -> Dict: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Dict , ) -> str: lowerCamelCase__ : Dict = UMTaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , ) lowerCamelCase__ : Any = model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) lowerCamelCase__ : Dict = result.last_hidden_state lowerCamelCase__ : Any = result.past_key_values lowerCamelCase__ : List[Any] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[Any] , ) -> Optional[int]: lowerCamelCase__ : List[Any] = UMTaModel(config=UpperCAmelCase ).get_decoder().to(UpperCAmelCase ).eval() # first forward pass lowerCamelCase__ : Tuple = model(UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , use_cache=UpperCAmelCase ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) + 1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCamelCase__ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )['last_hidden_state'] lowerCamelCase__ : str = model(UpperCAmelCase , past_key_values=UpperCAmelCase )['last_hidden_state'] # select random slice lowerCamelCase__ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase__ : Tuple = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase__ : List[str] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase , UpperCAmelCase , atol=1e-3 ) ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = UMTaModel(config=UpperCAmelCase ).to(UpperCAmelCase ).half().eval() lowerCamelCase__ : Optional[int] = model(**UpperCAmelCase )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCAmelCase ).any().item() ) @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCAmelCase__ = (UMTaForConditionalGeneration,) if is_torch_available() else () UpperCAmelCase__ = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = True UpperCAmelCase__ = True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCAmelCase__ = [0.8, 0.9] def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Tuple = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCAmelCase , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: lowerCamelCase__ : int = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Any = config_and_inputs[0] lowerCamelCase__ : Any = UMTaForConditionalGeneration(UpperCAmelCase ).eval() model.to(UpperCAmelCase ) lowerCamelCase__ : Tuple = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), } for attn_name, (name, mask) in zip(UpperCAmelCase , head_masking.items() ): lowerCamelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCamelCase__ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ) lowerCamelCase__ : Tuple = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase , return_dict_in_generate=UpperCAmelCase , **UpperCAmelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCamelCase__ : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=UpperCAmelCase ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=UpperCAmelCase , legacy=UpperCAmelCase ) lowerCamelCase__ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowerCamelCase__ : Tuple = tokenizer(UpperCAmelCase , return_tensors='pt' , padding=UpperCAmelCase ).input_ids # fmt: off lowerCamelCase__ : Any = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model.generate(input_ids.to(UpperCAmelCase ) ) lowerCamelCase__ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowerCamelCase__ : Union[str, Any] = tokenizer.batch_decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: if not numbers: return 0 if not isinstance(_UpperCAmelCase , (list, tuple) ) or not all( isinstance(_UpperCAmelCase , _UpperCAmelCase ) for number in numbers ): raise ValueError('numbers must be an iterable of integers' ) lowerCamelCase__ : str = numbers[0] for i in range(1 , len(_UpperCAmelCase ) ): # update the maximum and minimum subarray products lowerCamelCase__ : Tuple = numbers[i] if number < 0: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = min_till_now, max_till_now lowerCamelCase__ : Tuple = max(_UpperCAmelCase , max_till_now * number ) lowerCamelCase__ : Any = min(_UpperCAmelCase , min_till_now * number ) # update the maximum product found till now lowerCamelCase__ : Dict = max(_UpperCAmelCase , _UpperCAmelCase ) return max_prod
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase : Dict=3 , UpperCAmelCase : Union[str, Any]=3 , UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , UpperCAmelCase : List[str]=(64,) , UpperCAmelCase : Any=2 , UpperCAmelCase : List[Any]=32 , UpperCAmelCase : List[str]="silu" , UpperCAmelCase : Tuple=True , ) -> Tuple: super().__init__() lowerCamelCase__ : Optional[Any] = layers_per_block lowerCamelCase__ : Dict = torch.nn.Convad( UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : Union[str, Any] = nn.ModuleList([] ) # down lowerCamelCase__ : Tuple = block_out_channels[0] for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : List[Any] = output_channel lowerCamelCase__ : Dict = block_out_channels[i] lowerCamelCase__ : int = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : int = get_down_block( UpperCAmelCase , num_layers=self.layers_per_block , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=UpperCAmelCase , resnet_groups=UpperCAmelCase , attention_head_dim=UpperCAmelCase , temb_channels=UpperCAmelCase , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Tuple = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=UpperCAmelCase , temb_channels=UpperCAmelCase , ) # out lowerCamelCase__ : Optional[Any] = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=UpperCAmelCase , eps=1e-6 ) lowerCamelCase__ : List[Any] = nn.SiLU() lowerCamelCase__ : List[str] = 2 * out_channels if double_z else out_channels lowerCamelCase__ : Any = nn.Convad(block_out_channels[-1] , UpperCAmelCase , 3 , padding=1 ) lowerCamelCase__ : List[Any] = False def A_ ( self : Dict , UpperCAmelCase : Optional[int] ) -> Union[str, Any]: lowerCamelCase__ : Dict = x lowerCamelCase__ : Dict = self.conv_in(UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(UpperCAmelCase : Dict ): def custom_forward(*UpperCAmelCase : Any ): return module(*UpperCAmelCase ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: lowerCamelCase__ : Optional[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(UpperCAmelCase ) , UpperCAmelCase , use_reentrant=UpperCAmelCase ) # middle lowerCamelCase__ : Tuple = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCAmelCase , use_reentrant=UpperCAmelCase ) else: for down_block in self.down_blocks: lowerCamelCase__ : str = torch.utils.checkpoint.checkpoint(create_custom_forward(UpperCAmelCase ) , UpperCAmelCase ) # middle lowerCamelCase__ : Dict = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , UpperCAmelCase ) else: # down for down_block in self.down_blocks: lowerCamelCase__ : Optional[Any] = down_block(UpperCAmelCase ) # middle lowerCamelCase__ : List[str] = self.mid_block(UpperCAmelCase ) # post-process lowerCamelCase__ : str = self.conv_norm_out(UpperCAmelCase ) lowerCamelCase__ : List[Any] = self.conv_act(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.conv_out(UpperCAmelCase ) return sample class lowerCAmelCase ( nn.Module ): def __init__( self : Any , UpperCAmelCase : Optional[int]=3 , UpperCAmelCase : List[Any]=3 , UpperCAmelCase : int=("UpDecoderBlock2D",) , UpperCAmelCase : Optional[Any]=(64,) , UpperCAmelCase : str=2 , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : int="silu" , UpperCAmelCase : List[str]="group" , ) -> Any: super().__init__() lowerCamelCase__ : List[str] = layers_per_block lowerCamelCase__ : Optional[Any] = nn.Convad( UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase__ : Dict = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Dict = in_channels if norm_type == 'spatial' else None # mid lowerCamelCase__ : List[Any] = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=UpperCAmelCase , temb_channels=UpperCAmelCase , ) # up lowerCamelCase__ : Tuple = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Dict = reversed_block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Dict = output_channel lowerCamelCase__ : List[Any] = reversed_block_out_channels[i] lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Union[str, Any] = get_up_block( UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , prev_output_channel=UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=UpperCAmelCase , resnet_groups=UpperCAmelCase , attention_head_dim=UpperCAmelCase , temb_channels=UpperCAmelCase , resnet_time_scale_shift=UpperCAmelCase , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : str = output_channel # out if norm_type == "spatial": lowerCamelCase__ : Any = SpatialNorm(block_out_channels[0] , UpperCAmelCase ) else: lowerCamelCase__ : Optional[Any] = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=UpperCAmelCase , eps=1e-6 ) lowerCamelCase__ : Tuple = nn.SiLU() lowerCamelCase__ : Union[str, Any] = nn.Convad(block_out_channels[0] , UpperCAmelCase , 3 , padding=1 ) lowerCamelCase__ : Dict = False def A_ ( self : Tuple , UpperCAmelCase : Tuple , UpperCAmelCase : Any=None ) -> List[str]: lowerCamelCase__ : Any = z lowerCamelCase__ : Union[str, Any] = self.conv_in(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(UpperCAmelCase : Any ): def custom_forward(*UpperCAmelCase : Any ): return module(*UpperCAmelCase ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle lowerCamelCase__ : str = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCAmelCase , UpperCAmelCase , use_reentrant=UpperCAmelCase ) lowerCamelCase__ : Tuple = sample.to(UpperCAmelCase ) # up for up_block in self.up_blocks: lowerCamelCase__ : str = torch.utils.checkpoint.checkpoint( create_custom_forward(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase , use_reentrant=UpperCAmelCase ) else: # middle lowerCamelCase__ : str = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Dict = sample.to(UpperCAmelCase ) # up for up_block in self.up_blocks: lowerCamelCase__ : Tuple = torch.utils.checkpoint.checkpoint(create_custom_forward(UpperCAmelCase ) , UpperCAmelCase , UpperCAmelCase ) else: # middle lowerCamelCase__ : Any = self.mid_block(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = sample.to(UpperCAmelCase ) # up for up_block in self.up_blocks: lowerCamelCase__ : List[Any] = up_block(UpperCAmelCase , UpperCAmelCase ) # post-process if latent_embeds is None: lowerCamelCase__ : str = self.conv_norm_out(UpperCAmelCase ) else: lowerCamelCase__ : Optional[Any] = self.conv_norm_out(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : List[str] = self.conv_act(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.conv_out(UpperCAmelCase ) return sample class lowerCAmelCase ( nn.Module ): def __init__( self : Dict , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]="random" , UpperCAmelCase : Optional[int]=False , UpperCAmelCase : List[Any]=True ) -> Optional[int]: super().__init__() lowerCamelCase__ : int = n_e lowerCamelCase__ : Any = vq_embed_dim lowerCamelCase__ : Optional[int] = beta lowerCamelCase__ : Optional[Any] = legacy lowerCamelCase__ : Any = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) lowerCamelCase__ : Any = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) lowerCamelCase__ : str = self.used.shape[0] lowerCamelCase__ : Any = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": lowerCamelCase__ : List[Any] = self.re_embed lowerCamelCase__ : List[Any] = self.re_embed + 1 print( F"""Remapping {self.n_e} indices to {self.re_embed} indices. """ F"""Using {self.unknown_index} for unknown indices.""" ) else: lowerCamelCase__ : List[Any] = n_e lowerCamelCase__ : Optional[Any] = sane_index_shape def A_ ( self : Union[str, Any] , UpperCAmelCase : Any ) -> Dict: lowerCamelCase__ : Any = inds.shape assert len(UpperCAmelCase ) > 1 lowerCamelCase__ : str = inds.reshape(ishape[0] , -1 ) lowerCamelCase__ : str = self.used.to(UpperCAmelCase ) lowerCamelCase__ : Dict = (inds[:, :, None] == used[None, None, ...]).long() lowerCamelCase__ : Optional[int] = match.argmax(-1 ) lowerCamelCase__ : Dict = match.sum(2 ) < 1 if self.unknown_index == "random": lowerCamelCase__ : Optional[int] = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: lowerCamelCase__ : int = self.unknown_index return new.reshape(UpperCAmelCase ) def A_ ( self : List[str] , UpperCAmelCase : Tuple ) -> Optional[int]: lowerCamelCase__ : List[Any] = inds.shape assert len(UpperCAmelCase ) > 1 lowerCamelCase__ : List[Any] = inds.reshape(ishape[0] , -1 ) lowerCamelCase__ : str = self.used.to(UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token lowerCamelCase__ : Any = 0 # simply set to zero lowerCamelCase__ : Any = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , UpperCAmelCase ) return back.reshape(UpperCAmelCase ) def A_ ( self : Tuple , UpperCAmelCase : Union[str, Any] ) -> Optional[int]: # reshape z -> (batch, height, width, channel) and flatten lowerCamelCase__ : Tuple = z.permute(0 , 2 , 3 , 1 ).contiguous() lowerCamelCase__ : List[Any] = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z lowerCamelCase__ : List[str] = torch.argmin(torch.cdist(UpperCAmelCase , self.embedding.weight ) , dim=1 ) lowerCamelCase__ : Optional[Any] = self.embedding(UpperCAmelCase ).view(z.shape ) lowerCamelCase__ : int = None lowerCamelCase__ : Union[str, Any] = None # compute loss for embedding if not self.legacy: lowerCamelCase__ : Dict = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: lowerCamelCase__ : List[str] = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients lowerCamelCase__ : Union[str, Any] = z + (z_q - z).detach() # reshape back to match original input shape lowerCamelCase__ : Any = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: lowerCamelCase__ : int = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis lowerCamelCase__ : Union[str, Any] = self.remap_to_used(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: lowerCamelCase__ : Any = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def A_ ( self : int , UpperCAmelCase : Dict , UpperCAmelCase : List[Any] ) -> List[Any]: # shape specifying (batch, height, width, channel) if self.remap is not None: lowerCamelCase__ : str = indices.reshape(shape[0] , -1 ) # add batch axis lowerCamelCase__ : Optional[Any] = self.unmap_to_all(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = indices.reshape(-1 ) # flatten again # get quantized latent vectors lowerCamelCase__ : str = self.embedding(UpperCAmelCase ) if shape is not None: lowerCamelCase__ : List[str] = z_q.view(UpperCAmelCase ) # reshape back to match original input shape lowerCamelCase__ : Union[str, Any] = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Any , UpperCAmelCase : Tuple , UpperCAmelCase : List[Any]=False ) -> List[str]: lowerCamelCase__ : List[str] = parameters lowerCamelCase__ , lowerCamelCase__ : List[str] = torch.chunk(UpperCAmelCase , 2 , dim=1 ) lowerCamelCase__ : Union[str, Any] = torch.clamp(self.logvar , -3_0.0 , 2_0.0 ) lowerCamelCase__ : List[Any] = deterministic lowerCamelCase__ : Optional[int] = torch.exp(0.5 * self.logvar ) lowerCamelCase__ : Optional[Any] = torch.exp(self.logvar ) if self.deterministic: lowerCamelCase__ : str = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def A_ ( self : List[str] , UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: # make sure sample is on the same device as the parameters and has same dtype lowerCamelCase__ : Any = randn_tensor( self.mean.shape , generator=UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) lowerCamelCase__ : Optional[int] = self.mean + self.std * sample return x def A_ ( self : Dict , UpperCAmelCase : Optional[Any]=None ) -> Optional[Any]: if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def A_ ( self : str , UpperCAmelCase : int , UpperCAmelCase : List[Any]=[1, 2, 3] ) -> List[str]: if self.deterministic: return torch.Tensor([0.0] ) lowerCamelCase__ : int = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=UpperCAmelCase ) def A_ ( self : Optional[int] ) -> List[str]: return self.mean
50
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase ): @register_to_config def __init__( self : List[str] , UpperCAmelCase : int = 65536 , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 0 , UpperCAmelCase : str = "fourier" , UpperCAmelCase : bool = True , UpperCAmelCase : bool = False , UpperCAmelCase : float = 0.0 , UpperCAmelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase : str = None , UpperCAmelCase : Tuple[int] = (32, 32, 64) , UpperCAmelCase : str = None , UpperCAmelCase : int = 8 , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = False , ) -> List[Any]: super().__init__() lowerCamelCase__ : Optional[int] = sample_size # time if time_embedding_type == "fourier": lowerCamelCase__ : Optional[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase , log=UpperCAmelCase , flip_sin_to_cos=UpperCAmelCase ) lowerCamelCase__ : Any = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCamelCase__ : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase , downscale_freq_shift=UpperCAmelCase ) lowerCamelCase__ : Dict = block_out_channels[0] if use_timestep_embedding: lowerCamelCase__ : str = block_out_channels[0] * 4 lowerCamelCase__ : List[Any] = TimestepEmbedding( in_channels=UpperCAmelCase , time_embed_dim=UpperCAmelCase , act_fn=UpperCAmelCase , out_dim=block_out_channels[0] , ) lowerCamelCase__ : Any = nn.ModuleList([] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Optional[int] = None # down lowerCamelCase__ : Optional[int] = in_channels for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = output_channel lowerCamelCase__ : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Optional[int] = get_down_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Optional[int] = get_mid_block( UpperCAmelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase , add_downsample=UpperCAmelCase , ) # up lowerCamelCase__ : Optional[int] = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = reversed_block_out_channels[0] if out_block_type is None: lowerCamelCase__ : List[str] = out_channels else: lowerCamelCase__ : Any = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Tuple = output_channel lowerCamelCase__ : Union[str, Any] = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase ) - 1 else final_upsample_channels ) lowerCamelCase__ : List[str] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Dict = get_up_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : int = output_channel # out lowerCamelCase__ : int = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCamelCase__ : List[Any] = get_out_block( out_block_type=UpperCAmelCase , num_groups_out=UpperCAmelCase , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase , act_fn=UpperCAmelCase , fc_dim=block_out_channels[-1] // 4 , ) def A_ ( self : List[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Union[torch.Tensor, float, int] , UpperCAmelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: lowerCamelCase__ : Optional[Any] = timestep if not torch.is_tensor(UpperCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(sample.device ) lowerCamelCase__ : Optional[int] = self.time_proj(UpperCAmelCase ) if self.config.use_timestep_embedding: lowerCamelCase__ : str = self.time_mlp(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = timestep_embed[..., None] lowerCamelCase__ : str = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCamelCase__ : str = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCamelCase__ : str = () for downsample_block in self.down_blocks: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = downsample_block(hidden_states=UpperCAmelCase , temb=UpperCAmelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCamelCase__ : Optional[Any] = self.mid_block(UpperCAmelCase , UpperCAmelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCamelCase__ : Dict = down_block_res_samples[-1:] lowerCamelCase__ : Optional[Any] = down_block_res_samples[:-1] lowerCamelCase__ : Any = upsample_block(UpperCAmelCase , res_hidden_states_tuple=UpperCAmelCase , temb=UpperCAmelCase ) # 5. post-process if self.out_block: lowerCamelCase__ : Any = self.out_block(UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase )
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : int = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : List[str] = 0 while b > 0: if b & 1: lowerCamelCase__ : int = ((res % c) + (a % c)) % c a += a b >>= 1 return res
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata _UpperCAmelCase : str = """""" if version.parse(importlib_metadata.version("""jiwer""")) < version.parse("""2.3.0"""): class lowerCAmelCase ( tr.AbstractTransform ): def __init__( self : Tuple , UpperCAmelCase : str = " " ) -> Tuple: lowerCamelCase__ : Dict = sentence_delimiter def A_ ( self : Dict , UpperCAmelCase : str ) -> Any: return list(UpperCAmelCase ) def A_ ( self : Union[str, Any] , UpperCAmelCase : List[str] ) -> Tuple: lowerCamelCase__ : Tuple = [] for sent_idx, sentence in enumerate(UpperCAmelCase ): chars.extend(self.process_string(UpperCAmelCase ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(UpperCAmelCase ) - 1: chars.append(self.sentence_delimiter ) return chars _UpperCAmelCase : str = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: _UpperCAmelCase : Optional[Any] = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) _UpperCAmelCase : str = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ _UpperCAmelCase : List[Any] = """\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER's output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. """ _UpperCAmelCase : str = """ Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> cer = datasets.load_metric(\"cer\") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def A_ ( self : int ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/jitsi/jiwer/'] , reference_urls=[ 'https://en.wikipedia.org/wiki/Word_error_rate', 'https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates', ] , ) def A_ ( self : Tuple , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=False ) -> Any: if concatenate_texts: return jiwer.compute_measures( UpperCAmelCase , UpperCAmelCase , truth_transform=UpperCAmelCase , hypothesis_transform=UpperCAmelCase , )["wer"] lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : List[Any] = 0 for prediction, reference in zip(UpperCAmelCase , UpperCAmelCase ): lowerCamelCase__ : Optional[int] = jiwer.compute_measures( UpperCAmelCase , UpperCAmelCase , truth_transform=UpperCAmelCase , hypothesis_transform=UpperCAmelCase , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
50
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[int] = [] embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", F"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", F"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", F"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", F"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Tuple = [] attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", F"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", F"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", F"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", F"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] token.append((F"""cvt.encoder.stages.{idx}.cls_token""", 'stage2.cls_token') ) return token def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Tuple = 'imagenet-1k-id2label.json' lowerCamelCase__ : Union[str, Any] = 1000 lowerCamelCase__ : Optional[Any] = 'huggingface/label-files' lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowerCamelCase__ : int = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : List[str] = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowerCamelCase__ : List[Any] = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowerCamelCase__ : Dict = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowerCamelCase__ : Optional[Any] = [2, 2, 20] lowerCamelCase__ : Optional[int] = [3, 12, 16] lowerCamelCase__ : str = [192, 768, 1024] lowerCamelCase__ : Any = CvtForImageClassification(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowerCamelCase__ : Optional[int] = OrderedDict() lowerCamelCase__ : Tuple = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowerCamelCase__ : Optional[Any] = list_of_state_dict + cls_token(_UpperCAmelCase ) lowerCamelCase__ : str = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowerCamelCase__ : str = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : int = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : str = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( """--cvt_model""", default="""cvt-w24""", type=str, help="""Name of the cvt model you'd like to convert.""", ) parser.add_argument( """--image_size""", default=3_84, type=int, help="""Input Image Size""", ) parser.add_argument( """--cvt_file_name""", default=R"""cvtmodels\CvT-w24-384x384-IN-22k.pth""", type=str, help="""Input Image Size""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : int = 0 lowerCamelCase__ : Tuple = len(_UpperCAmelCase ) for i in range(n - 1 ): for j in range(i + 1 , _UpperCAmelCase ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: if len(_UpperCAmelCase ) <= 1: return arr, 0 lowerCamelCase__ : Tuple = len(_UpperCAmelCase ) // 2 lowerCamelCase__ : Union[str, Any] = arr[0:mid] lowerCamelCase__ : List[Any] = arr[mid:] lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = count_inversions_recursive(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : str = count_inversions_recursive(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = _count_cross_inversions(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Any = inversion_p + inversions_q + cross_inversions return c, num_inversions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] lowerCamelCase__ : Tuple = 0 while i < len(_UpperCAmelCase ) and j < len(_UpperCAmelCase ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(_UpperCAmelCase ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(_UpperCAmelCase ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowerCamelCase__ : Dict = count_inversions_bf(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : int = count_inversions_recursive(_UpperCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 8 print('number of inversions = ' , _UpperCAmelCase ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowerCamelCase__ : Optional[Any] = count_inversions_bf(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = count_inversions_recursive(_UpperCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , _UpperCAmelCase ) # an empty list should also have zero inversions lowerCamelCase__ : Optional[int] = [] lowerCamelCase__ : str = count_inversions_bf(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Any = count_inversions_recursive(_UpperCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , _UpperCAmelCase ) if __name__ == "__main__": main()
50
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
1
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever _UpperCAmelCase : Union[str, Any] = logging.getLogger(__name__) class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Dict , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Any , UpperCAmelCase : Any=None ) -> Any: super().__init__( UpperCAmelCase , question_encoder_tokenizer=UpperCAmelCase , generator_tokenizer=UpperCAmelCase , index=UpperCAmelCase , init_retrieval=UpperCAmelCase , ) lowerCamelCase__ : Union[str, Any] = None def A_ ( self : Optional[Any] , UpperCAmelCase : int ) -> Optional[Any]: logger.info('initializing retrieval' ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info('dist initialized' ) # needs to be set manually lowerCamelCase__ : Dict = self._infer_socket_ifname() # avoid clash with the NCCL port lowerCamelCase__ : Union[str, Any] = str(distributed_port + 1 ) lowerCamelCase__ : int = dist.new_group(ranks=UpperCAmelCase , backend='gloo' ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info('dist not initialized / main' ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def A_ ( self : List[Any] ) -> List[Any]: return dist.get_rank(group=self.process_group ) == 0 def A_ ( self : Dict , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str]=torch.floataa ) -> List[Any]: lowerCamelCase__ : Dict = torch.empty(UpperCAmelCase , dtype=UpperCAmelCase ) dist.scatter(UpperCAmelCase , src=0 , scatter_list=UpperCAmelCase , group=self.process_group ) return target_tensor def A_ ( self : str ) -> Union[str, Any]: lowerCamelCase__ : Any = psutil.net_if_addrs() # a hacky way to deal with varying network interface names lowerCamelCase__ : Any = next((addr for addr in addrs if addr.startswith('e' )) , UpperCAmelCase ) return ifname def A_ ( self : Dict , UpperCAmelCase : np.ndarray , UpperCAmelCase : int ) -> Tuple[np.ndarray, List[dict]]: # single GPU training if not dist.is_initialized(): lowerCamelCase__ , lowerCamelCase__ : int = self._main_retrieve(UpperCAmelCase , UpperCAmelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(UpperCAmelCase ) # distributed training lowerCamelCase__ : List[Any] = dist.get_world_size(group=self.process_group ) # gather logic lowerCamelCase__ : Any = None if self._is_main(): lowerCamelCase__ : List[str] = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(UpperCAmelCase )] dist.gather(torch.tensor(UpperCAmelCase ) , dst=0 , gather_list=UpperCAmelCase , group=self.process_group ) # scatter logic lowerCamelCase__ : List[str] = question_hidden_states.shape[0] lowerCamelCase__ : Union[str, Any] = [] lowerCamelCase__ : Union[str, Any] = [] if self._is_main(): assert len(UpperCAmelCase ) == world_size lowerCamelCase__ , lowerCamelCase__ : int = self._main_retrieve(torch.cat(UpperCAmelCase ).numpy() , UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = torch.tensor(UpperCAmelCase ), torch.tensor(UpperCAmelCase ) lowerCamelCase__ : str = self._chunk_tensor(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Dict = self._chunk_tensor(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Dict = self._scattered(UpperCAmelCase , [n_queries, n_docs] , target_type=torch.intaa ) lowerCamelCase__ : List[Any] = self._scattered(UpperCAmelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(UpperCAmelCase )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
1
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( """stable diffusion controlnet""", """0.22.0""", """Importing `StableDiffusionControlNetPipeline` or `MultiControlNetModel` from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import StableDiffusionControlNetPipeline` instead.""", standard_warn=False, stacklevel=3, )
50
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } _UpperCAmelCase : Any = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]="[UNK]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Tuple="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=None , **UpperCAmelCase : Optional[int] , ) -> str: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : str = strip_accents lowerCamelCase__ : Optional[Any] = tokenize_chinese_chars lowerCamelCase__ : int = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : str = do_lower_case def A_ ( self : Optional[int] , UpperCAmelCase : int , **UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH lowerCamelCase__ : Optional[int] = text lowerCamelCase__ : Dict = kwargs.pop('text_pair' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = kwargs.pop('return_tensors' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: lowerCamelCase__ : Tuple = batch_text_pair[idx] else: lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Any = encoded_candidates.get('input_ids' ) lowerCamelCase__ : Union[str, Any] = encoded_candidates.get('attention_mask' ) lowerCamelCase__ : Tuple = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) lowerCamelCase__ : int = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ) -> List[str]: lowerCamelCase__ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
1
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowerCAmelCase : def __init__( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Any=2 , UpperCAmelCase : Dict=3 , UpperCAmelCase : List[str]=4 , UpperCAmelCase : str=2 , UpperCAmelCase : int=7 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : str=True , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : int=99 , UpperCAmelCase : List[Any]=36 , UpperCAmelCase : Dict=2 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Optional[Any]=37 , UpperCAmelCase : str="gelu" , UpperCAmelCase : str=0.1 , UpperCAmelCase : Tuple=0.1 , UpperCAmelCase : Tuple=512 , UpperCAmelCase : Optional[int]=16 , UpperCAmelCase : int=2 , UpperCAmelCase : Union[str, Any]=0.0_2 , UpperCAmelCase : Optional[int]=6 , UpperCAmelCase : List[Any]=6 , UpperCAmelCase : Tuple=3 , UpperCAmelCase : Optional[int]=4 , UpperCAmelCase : List[str]=None , UpperCAmelCase : Optional[int]=1000 , ) -> Optional[int]: lowerCamelCase__ : Optional[int] = parent lowerCamelCase__ : Tuple = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : str = image_size lowerCamelCase__ : List[str] = patch_size lowerCamelCase__ : str = is_training lowerCamelCase__ : int = use_input_mask lowerCamelCase__ : Union[str, Any] = use_token_type_ids lowerCamelCase__ : Any = use_labels lowerCamelCase__ : Dict = vocab_size lowerCamelCase__ : Any = hidden_size lowerCamelCase__ : List[str] = num_hidden_layers lowerCamelCase__ : Union[str, Any] = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : str = hidden_act lowerCamelCase__ : List[Any] = hidden_dropout_prob lowerCamelCase__ : Any = attention_probs_dropout_prob lowerCamelCase__ : List[str] = max_position_embeddings lowerCamelCase__ : List[Any] = type_vocab_size lowerCamelCase__ : Any = type_sequence_label_size lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Tuple = coordinate_size lowerCamelCase__ : Any = shape_size lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : List[str] = num_choices lowerCamelCase__ : Tuple = scope lowerCamelCase__ : List[Any] = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) lowerCamelCase__ : List[Any] = text_seq_length lowerCamelCase__ : Tuple = (image_size // patch_size) ** 2 + 1 lowerCamelCase__ : Optional[Any] = self.text_seq_length + self.image_seq_length def A_ ( self : Union[str, Any] ) -> Any: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) lowerCamelCase__ : Union[str, Any] = bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCamelCase__ : str = bbox[i, j, 3] lowerCamelCase__ : Union[str, Any] = bbox[i, j, 1] lowerCamelCase__ : Optional[int] = tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: lowerCamelCase__ : Optional[int] = bbox[i, j, 2] lowerCamelCase__ : Optional[int] = bbox[i, j, 0] lowerCamelCase__ : Any = tmp_coordinate lowerCamelCase__ : List[str] = tf.constant(UpperCAmelCase ) lowerCamelCase__ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None if self.use_input_mask: lowerCamelCase__ : int = random_attention_mask([self.batch_size, self.text_seq_length] ) lowerCamelCase__ : int = None if self.use_token_type_ids: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) lowerCamelCase__ : Dict = None lowerCamelCase__ : Any = None if self.use_labels: lowerCamelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) lowerCamelCase__ : int = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def A_ ( self : List[str] , UpperCAmelCase : List[str] , UpperCAmelCase : Any , UpperCAmelCase : List[Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Any ) -> str: lowerCamelCase__ : List[Any] = TFLayoutLMvaModel(config=UpperCAmelCase ) # text + image lowerCamelCase__ : str = model(UpperCAmelCase , pixel_values=UpperCAmelCase , training=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model( UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , training=UpperCAmelCase , ) lowerCamelCase__ : Any = model(UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , training=UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only lowerCamelCase__ : Optional[int] = model(UpperCAmelCase , training=UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only lowerCamelCase__ : int = model({'pixel_values': pixel_values} , training=UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def A_ ( self : Tuple , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict , UpperCAmelCase : Any , UpperCAmelCase : str , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[Any] ) -> Optional[int]: lowerCamelCase__ : Optional[int] = self.num_labels lowerCamelCase__ : Tuple = TFLayoutLMvaForSequenceClassification(config=UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = model( UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , labels=UpperCAmelCase , training=UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A_ ( self : Tuple , UpperCAmelCase : str , UpperCAmelCase : Tuple , UpperCAmelCase : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Any , UpperCAmelCase : List[str] ) -> str: lowerCamelCase__ : List[Any] = self.num_labels lowerCamelCase__ : Any = TFLayoutLMvaForTokenClassification(config=UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model( UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , labels=UpperCAmelCase , training=UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def A_ ( self : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int] , UpperCAmelCase : List[str] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Optional[int] ) -> Optional[Any]: lowerCamelCase__ : List[Any] = 2 lowerCamelCase__ : List[str] = TFLayoutLMvaForQuestionAnswering(config=UpperCAmelCase ) lowerCamelCase__ : Any = model( UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , attention_mask=UpperCAmelCase , token_type_ids=UpperCAmelCase , start_positions=UpperCAmelCase , end_positions=UpperCAmelCase , training=UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A_ ( self : str ) -> Optional[Any]: lowerCamelCase__ : Tuple = self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Union[str, Any] = config_and_inputs lowerCamelCase__ : Dict = { 'input_ids': input_ids, 'bbox': bbox, 'pixel_values': pixel_values, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_tf class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) UpperCAmelCase__ = ( {"""document-question-answering""": TFLayoutLMvaForQuestionAnswering, """feature-extraction""": TFLayoutLMvaModel} if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : Tuple , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[int] , UpperCAmelCase : int , UpperCAmelCase : Union[str, Any] ) -> str: return True def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Optional[int]=False ) -> dict: lowerCamelCase__ : str = copy.deepcopy(UpperCAmelCase ) if model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : int = { k: tf.tile(tf.expand_dims(UpperCAmelCase , 1 ) , (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(UpperCAmelCase , tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : List[Any] = tf.ones(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : int = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) lowerCamelCase__ : str = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : Tuple = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=tf.intaa ) return inputs_dict def A_ ( self : Dict ) -> Union[str, Any]: lowerCamelCase__ : str = TFLayoutLMvaModelTester(self ) lowerCamelCase__ : int = ConfigTester(self , config_class=UpperCAmelCase , hidden_size=37 ) def A_ ( self : Union[str, Any] ) -> Tuple: self.config_tester.run_common_tests() def A_ ( self : Optional[int] ) -> Tuple: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[Any] = model_class(UpperCAmelCase ) if getattr(UpperCAmelCase , 'hf_compute_loss' , UpperCAmelCase ): # The number of elements in the loss should be the same as the number of elements in the label lowerCamelCase__ : Optional[Any] = self._prepare_for_class(inputs_dict.copy() , UpperCAmelCase , return_labels=UpperCAmelCase ) lowerCamelCase__ : Any = prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys() , reverse=UpperCAmelCase )[0] ] lowerCamelCase__ : List[str] = added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs lowerCamelCase__ : str = self._prepare_for_class(inputs_dict.copy() , UpperCAmelCase , return_labels=UpperCAmelCase ) lowerCamelCase__ : Tuple = prepared_for_class.pop('input_ids' ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase , **UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions lowerCamelCase__ : Any = self._prepare_for_class(inputs_dict.copy() , UpperCAmelCase , return_labels=UpperCAmelCase ) lowerCamelCase__ : Tuple = prepared_for_class.pop('input_ids' ) if "labels" in prepared_for_class: lowerCamelCase__ : Tuple = prepared_for_class['labels'].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: lowerCamelCase__ : Union[str, Any] = -100 lowerCamelCase__ : List[str] = tf.convert_to_tensor(UpperCAmelCase ) lowerCamelCase__ : Tuple = model(UpperCAmelCase , **UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict lowerCamelCase__ : Any = self._prepare_for_class(inputs_dict.copy() , UpperCAmelCase , return_labels=UpperCAmelCase ) lowerCamelCase__ : Any = model(UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple lowerCamelCase__ : Dict = self._prepare_for_class(inputs_dict.copy() , UpperCAmelCase , return_labels=UpperCAmelCase ) # Get keys that were added with the _prepare_for_class function lowerCamelCase__ : Dict = prepared_for_class.keys() - inputs_dict.keys() lowerCamelCase__ : Tuple = inspect.signature(model.call ).parameters lowerCamelCase__ : str = list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple lowerCamelCase__ : Optional[Any] = {0: 'input_ids'} for label_key in label_keys: lowerCamelCase__ : int = signature_names.index(UpperCAmelCase ) lowerCamelCase__ : int = label_key lowerCamelCase__ : int = sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple lowerCamelCase__ : Tuple = [] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: lowerCamelCase__ : int = prepared_for_class[value] lowerCamelCase__ : Union[str, Any] = tuple(UpperCAmelCase ) # Send to model lowerCamelCase__ : Tuple = model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def A_ ( self : List[str] ) -> List[Any]: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase__ : Tuple = type self.model_tester.create_and_check_model(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Any: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : str ) -> List[str]: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) @slow def A_ ( self : List[Any] ) -> List[Any]: for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : List[Any] = TFLayoutLMvaModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf class lowerCAmelCase ( unittest.TestCase ): @cached_property def A_ ( self : int ) -> Tuple: return LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase ) if is_vision_available() else None @slow def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : List[Any] = TFLayoutLMvaModel.from_pretrained('microsoft/layoutlmv3-base' ) lowerCamelCase__ : List[str] = self.default_image_processor lowerCamelCase__ : Any = prepare_img() lowerCamelCase__ : Any = image_processor(images=UpperCAmelCase , return_tensors='tf' ).pixel_values lowerCamelCase__ : List[Any] = tf.constant([[1, 2]] ) lowerCamelCase__ : Tuple = tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ) , axis=0 ) # forward pass lowerCamelCase__ : Optional[int] = model(input_ids=UpperCAmelCase , bbox=UpperCAmelCase , pixel_values=UpperCAmelCase , training=UpperCAmelCase ) # verify the logits lowerCamelCase__ : str = (1, 199, 768) self.assertEqual(outputs.last_hidden_state.shape , UpperCAmelCase ) lowerCamelCase__ : List[str] = tf.constant( [[-0.0_5_2_9, 0.3_6_1_8, 0.1_6_3_2], [-0.1_5_8_7, -0.1_6_6_7, -0.0_4_0_0], [-0.1_5_5_7, -0.1_6_7_1, -0.0_5_0_5]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
50
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _UpperCAmelCase : Optional[Any] = logging.getLogger(__name__) @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if is_torch_available(): import torch from torch.utils.data import Dataset class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 def __init__( self : int , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : List[str]=False , UpperCAmelCase : bool = False , ) -> List[str]: lowerCamelCase__ : int = hans_processors[task]() lowerCamelCase__ : Optional[Any] = os.path.join( UpperCAmelCase , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase ) , UpperCAmelCase , ) , ) lowerCamelCase__ : int = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = label_list[2], label_list[1] lowerCamelCase__ : List[str] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : str = cached_features_file + '.lock' with FileLock(UpperCAmelCase ): if os.path.exists(UpperCAmelCase ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) lowerCamelCase__ : int = torch.load(UpperCAmelCase ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) lowerCamelCase__ : str = ( processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) ) logger.info('Training examples: %s' , len(UpperCAmelCase ) ) lowerCamelCase__ : Dict = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) logger.info('Saving features into cached file %s' , UpperCAmelCase ) torch.save(self.features , UpperCAmelCase ) def __len__( self : Optional[int] ) -> Optional[Any]: return len(self.features ) def __getitem__( self : Tuple , UpperCAmelCase : Dict ) -> InputFeatures: return self.features[i] def A_ ( self : int ) -> int: return self.label_list if is_tf_available(): import tensorflow as tf class lowerCAmelCase : UpperCAmelCase__ = 42 def __init__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = 128 , UpperCAmelCase : Any=False , UpperCAmelCase : bool = False , ) -> Union[str, Any]: lowerCamelCase__ : Any = hans_processors[task]() lowerCamelCase__ : Optional[Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : str = label_list[2], label_list[1] lowerCamelCase__ : Optional[int] = label_list lowerCamelCase__ : int = processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCamelCase__ : Optional[int] = tf.data.Dataset.from_generator( UpperCAmelCase , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def A_ ( self : Any ) -> Any: return self.dataset def __len__( self : Tuple ) -> int: return len(self.features ) def __getitem__( self : List[str] , UpperCAmelCase : Any ) -> InputFeatures: return self.features[i] def A_ ( self : Dict ) -> str: return self.label_list class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : int , UpperCAmelCase : List[Any] ) -> int: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_train_set.txt' ) ) , 'train' ) def A_ ( self : Any , UpperCAmelCase : int ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def A_ ( self : Any ) -> List[Any]: return ["contradiction", "entailment", "neutral"] def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> List[str]: lowerCamelCase__ : List[str] = [] for i, line in enumerate(UpperCAmelCase ): if i == 0: continue lowerCamelCase__ : Tuple = '%s-%s' % (set_type, line[0]) lowerCamelCase__ : str = line[5] lowerCamelCase__ : Dict = line[6] lowerCamelCase__ : int = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCamelCase__ : Dict = line[0] examples.append(InputExample(guid=UpperCAmelCase , text_a=UpperCAmelCase , text_b=UpperCAmelCase , label=UpperCAmelCase , pairID=UpperCAmelCase ) ) return examples def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> Optional[int]: lowerCamelCase__ : int = {label: i for i, label in enumerate(_UpperCAmelCase )} lowerCamelCase__ : List[Any] = [] for ex_index, example in tqdm.tqdm(enumerate(_UpperCAmelCase ) , desc='convert examples to features' ): if ex_index % 1_0000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCamelCase__ : List[Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=_UpperCAmelCase , max_length=_UpperCAmelCase , padding='max_length' , truncation=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , ) lowerCamelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCamelCase__ : Optional[int] = int(example.pairID ) features.append(InputFeatures(**_UpperCAmelCase , label=_UpperCAmelCase , pairID=_UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features _UpperCAmelCase : str = { """hans""": 3, } _UpperCAmelCase : List[Any] = { """hans""": HansProcessor, }
50
1
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[Any] = """▁""" _UpperCAmelCase : Optional[Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = BertGenerationTokenizer UpperCAmelCase__ = False UpperCAmelCase__ = True def A_ ( self : List[Any] ) -> List[str]: super().setUp() lowerCamelCase__ : Dict = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : List[str] = '<s>' lowerCamelCase__ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : List[str] ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(UpperCAmelCase ) , 1002 ) def A_ ( self : List[Any] ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowerCamelCase__ : List[str] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowerCamelCase__ : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def A_ ( self : Dict ) -> Tuple: return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def A_ ( self : Optional[int] ) -> List[str]: lowerCamelCase__ : Union[str, Any] = 'Hello World!' lowerCamelCase__ : Dict = [18536, 2260, 101] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @slow def A_ ( self : Optional[Any] ) -> str: lowerCamelCase__ : List[Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCamelCase__ : Any = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @require_torch @slow def A_ ( self : int ) -> Optional[Any]: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowerCamelCase__ : str = list(self.big_tokenizer.get_vocab().keys() )[:10] lowerCamelCase__ : int = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.big_tokenizer.encode_plus(UpperCAmelCase , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Tuple = BertGenerationConfig() lowerCamelCase__ : Optional[Any] = BertGenerationEncoder(UpperCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase ) model(**UpperCAmelCase ) @slow def A_ ( self : Optional[int] ) -> List[Any]: # fmt: off lowerCamelCase__ : Any = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
50
1
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _UpperCAmelCase : str = pytest.mark.integration @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : List[Any] ) -> Union[str, Any]: lowerCamelCase__ : List[Any] = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(UpperCAmelCase ) for x in np.arange(30 ).tolist()]} ) return dset def A_ ( self : Optional[Any] ) -> Optional[int]: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() lowerCamelCase__ : List[Any] = dset.map( lambda UpperCAmelCase , UpperCAmelCase : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=UpperCAmelCase , keep_in_memory=UpperCAmelCase ) lowerCamelCase__ : Tuple = dset.add_faiss_index('vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def A_ ( self : Union[str, Any] ) -> int: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : List[str] ) -> Tuple: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : Any ) -> Optional[Any]: lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(UpperCAmelCase , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def A_ ( self : Dict ) -> Dict: from elasticsearch import Elasticsearch lowerCamelCase__ : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : List[Any] = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) lowerCamelCase__ : int = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} lowerCamelCase__ : List[str] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Any ) -> Dict: import faiss lowerCamelCase__ : Tuple = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query lowerCamelCase__ : int = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Any = 1 lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = index.search(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries lowerCamelCase__ : str = np.eye(5 , dtype=np.floataa )[::-1] lowerCamelCase__ , lowerCamelCase__ : Dict = index.search_batch(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search_batch , queries[0] ) lowerCamelCase__ : str = [scores[0] for scores in total_scores] lowerCamelCase__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , UpperCAmelCase ) def A_ ( self : List[Any] ) -> List[Any]: import faiss lowerCamelCase__ : Any = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) lowerCamelCase__ : Tuple = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(UpperCAmelCase ): lowerCamelCase__ : List[str] = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def A_ ( self : List[str] ) -> Optional[int]: import faiss lowerCamelCase__ : Optional[Any] = faiss.IndexFlat(5 ) lowerCamelCase__ : int = FaissIndex(custom_index=UpperCAmelCase ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def A_ ( self : Any ) -> Optional[int]: import faiss lowerCamelCase__ : Any = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: index.save(tmp_file.name ) lowerCamelCase__ : List[Any] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ : List[str] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Tuple = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(UpperCAmelCase ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: import faiss lowerCamelCase__ : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) lowerCamelCase__ : Optional[int] = 'index.faiss' lowerCamelCase__ : Optional[Any] = F"""mock://{index_name}""" index.save(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Tuple = FaissIndex.load(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Optional[int] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Dict = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(_UpperCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Dict ) -> List[Any]: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : Any = Elasticsearch() lowerCamelCase__ : Tuple = {'acknowledged': True} lowerCamelCase__ : Tuple = ElasticSearchIndex(es_client=UpperCAmelCase ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query lowerCamelCase__ : Optional[int] = 'foo' lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search(UpperCAmelCase ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout lowerCamelCase__ : Any = 'foo' lowerCamelCase__ : List[str] = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Tuple = index.search(UpperCAmelCase , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries lowerCamelCase__ : List[str] = ['foo', 'bar', 'foobar'] lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : str = index.search_batch(UpperCAmelCase ) lowerCamelCase__ : List[str] = [scores[0] for scores in total_scores] lowerCamelCase__ : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase ) # batched queries with timeout lowerCamelCase__ : str = ['foo', 'bar', 'foobar'] lowerCamelCase__ : List[Any] = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search_batch(UpperCAmelCase , request_timeout=30 ) lowerCamelCase__ : Optional[Any] = [scores[0] for scores in total_scores] lowerCamelCase__ : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase )
50
1
from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase__ = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int] , UpperCAmelCase : str=False ) -> int: lowerCamelCase__ : Dict = super()._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) if return_labels: if model_class in get_values(UpperCAmelCase ): lowerCamelCase__ : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Optional[Any]=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : str=True , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : Tuple=32 , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : Union[str, Any]=2 , UpperCAmelCase : Dict=4 , UpperCAmelCase : Tuple=37 , UpperCAmelCase : Optional[int]="gelu" , UpperCAmelCase : int=0.1 , UpperCAmelCase : Any=0.1 , UpperCAmelCase : Optional[int]=512 , UpperCAmelCase : str=16 , UpperCAmelCase : int=2 , UpperCAmelCase : Dict=0.0_2 , UpperCAmelCase : Union[str, Any]=3 , UpperCAmelCase : str=4 , UpperCAmelCase : List[str]=None , ) -> Tuple: lowerCamelCase__ : List[str] = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : List[str] = seq_length lowerCamelCase__ : List[str] = is_training lowerCamelCase__ : Optional[Any] = use_input_mask lowerCamelCase__ : List[Any] = use_token_type_ids lowerCamelCase__ : Union[str, Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : List[Any] = hidden_size lowerCamelCase__ : int = num_hidden_layers lowerCamelCase__ : Union[str, Any] = num_attention_heads lowerCamelCase__ : List[Any] = intermediate_size lowerCamelCase__ : Tuple = hidden_act lowerCamelCase__ : int = hidden_dropout_prob lowerCamelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCamelCase__ : Tuple = max_position_embeddings lowerCamelCase__ : Optional[int] = type_vocab_size lowerCamelCase__ : Any = type_sequence_label_size lowerCamelCase__ : str = initializer_range lowerCamelCase__ : Dict = num_labels lowerCamelCase__ : List[str] = num_choices lowerCamelCase__ : Optional[Any] = scope lowerCamelCase__ : Dict = embedding_size def A_ ( self : Union[str, Any] ) -> Tuple: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : int = None if self.use_input_mask: lowerCamelCase__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase__ : List[str] = None if self.use_token_type_ids: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase__ : Optional[int] = None lowerCamelCase__ : Union[str, Any] = None lowerCamelCase__ : Optional[int] = None if self.use_labels: lowerCamelCase__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase__ : Optional[Any] = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A_ ( self : str , UpperCAmelCase : Tuple , UpperCAmelCase : List[str] , UpperCAmelCase : List[str] , UpperCAmelCase : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : Dict , UpperCAmelCase : Tuple ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = TFMobileBertModel(config=UpperCAmelCase ) lowerCamelCase__ : Tuple = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : Dict = model(UpperCAmelCase ) lowerCamelCase__ : List[Any] = [input_ids, input_mask] lowerCamelCase__ : Optional[Any] = model(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = model(UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A_ ( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Tuple , UpperCAmelCase : str , UpperCAmelCase : Tuple , UpperCAmelCase : Optional[Any] ) -> Any: lowerCamelCase__ : List[Any] = TFMobileBertForMaskedLM(config=UpperCAmelCase ) lowerCamelCase__ : str = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : Tuple = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A_ ( self : List[Any] , UpperCAmelCase : Optional[int] , UpperCAmelCase : Any , UpperCAmelCase : List[str] , UpperCAmelCase : Tuple , UpperCAmelCase : Any , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[Any] ) -> str: lowerCamelCase__ : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=UpperCAmelCase ) lowerCamelCase__ : Optional[int] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : str = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def A_ ( self : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : List[str] , UpperCAmelCase : List[str] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = TFMobileBertForPreTraining(config=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : Dict = model(UpperCAmelCase ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def A_ ( self : Tuple , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : Tuple , UpperCAmelCase : int , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : Optional[Any] ) -> Optional[Any]: lowerCamelCase__ : List[Any] = self.num_labels lowerCamelCase__ : Tuple = TFMobileBertForSequenceClassification(config=UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : Optional[Any] = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A_ ( self : Dict , UpperCAmelCase : Any , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[int] , UpperCAmelCase : Dict , UpperCAmelCase : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : List[Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = self.num_choices lowerCamelCase__ : List[Any] = TFMobileBertForMultipleChoice(config=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = tf.tile(tf.expand_dims(UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase__ : Optional[Any] = tf.tile(tf.expand_dims(UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase__ : Tuple = tf.tile(tf.expand_dims(UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase__ : int = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } lowerCamelCase__ : Optional[int] = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A_ ( self : Dict , UpperCAmelCase : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : str , UpperCAmelCase : int , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Dict , UpperCAmelCase : int ) -> int: lowerCamelCase__ : Optional[Any] = self.num_labels lowerCamelCase__ : List[str] = TFMobileBertForTokenClassification(config=UpperCAmelCase ) lowerCamelCase__ : List[Any] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : Optional[Any] = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A_ ( self : List[str] , UpperCAmelCase : Any , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Tuple , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Any , UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[Any] = TFMobileBertForQuestionAnswering(config=UpperCAmelCase ) lowerCamelCase__ : List[str] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A_ ( self : Optional[int] ) -> List[str]: lowerCamelCase__ : Optional[int] = self.prepare_config_and_inputs() ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : int = config_and_inputs lowerCamelCase__ : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict def A_ ( self : str ) -> Optional[Any]: lowerCamelCase__ : str = TFMobileBertModelTest.TFMobileBertModelTester(self ) lowerCamelCase__ : Optional[Any] = ConfigTester(self , config_class=UpperCAmelCase , hidden_size=37 ) def A_ ( self : Tuple ) -> List[str]: self.config_tester.run_common_tests() def A_ ( self : Union[str, Any] ) -> Tuple: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Union[str, Any]: lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*UpperCAmelCase ) def A_ ( self : Optional[int] ) -> Any: lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*UpperCAmelCase ) def A_ ( self : int ) -> Optional[Any]: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*UpperCAmelCase ) def A_ ( self : Tuple ) -> Optional[int]: lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Any: lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Dict: lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*UpperCAmelCase ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*UpperCAmelCase ) @slow def A_ ( self : List[str] ) -> List[Any]: # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: lowerCamelCase__ : Union[str, Any] = TFMobileBertModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @require_tf class lowerCAmelCase ( unittest.TestCase ): @slow def A_ ( self : Optional[Any] ) -> Union[str, Any]: lowerCamelCase__ : int = TFMobileBertForPreTraining.from_pretrained('google/mobilebert-uncased' ) lowerCamelCase__ : List[str] = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )[0] lowerCamelCase__ : Tuple = [1, 6, 30522] self.assertEqual(output.shape , UpperCAmelCase ) lowerCamelCase__ : List[Any] = tf.constant( [ [ [-4.5_9_1_9_5_4_7, -9.2_4_8_2_9_5, -9.6_4_5_2_5_6], [-6.7_3_0_6_1_7_5, -6.4_4_0_2_8_4, -6.6_0_5_2_8_3_7], [-7.2_7_4_3_5_0_6, -6.7_8_4_7_9_1_5, -6.0_2_4_6_7_3], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , UpperCAmelCase , atol=1e-4 )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : str = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCamelCase__ : Tuple = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCamelCase__ : Dict = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCamelCase__ : str = subset[i - 1][j] if arr[i - 1] <= j: lowerCamelCase__ : Dict = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
50
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _UpperCAmelCase : Tuple = 16 _UpperCAmelCase : Optional[int] = 32 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase = 16 ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCamelCase__ : List[str] = load_dataset('glue' , 'mrpc' ) def tokenize_function(_UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase__ : Optional[Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase__ : Dict = datasets.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase__ : Tuple = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase__ : Optional[Any] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase__ : int = 16 elif accelerator.mixed_precision != "no": lowerCamelCase__ : Any = 8 else: lowerCamelCase__ : int = None return tokenizer.pad( _UpperCAmelCase , padding='longest' , max_length=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_tensors='pt' , ) # Instantiate dataloaders. lowerCamelCase__ : Optional[int] = DataLoader( tokenized_datasets['train'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _UpperCAmelCase : str = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , _UpperCAmelCase ) == "1": lowerCamelCase__ : Optional[int] = 2 # New Code # lowerCamelCase__ : int = int(args.gradient_accumulation_steps ) # Initialize accelerator lowerCamelCase__ : Dict = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_UpperCAmelCase ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( 'Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase__ : Any = config['lr'] lowerCamelCase__ : Union[str, Any] = int(config['num_epochs'] ) lowerCamelCase__ : Optional[int] = int(config['seed'] ) lowerCamelCase__ : List[Any] = int(config['batch_size'] ) lowerCamelCase__ : Optional[Any] = evaluate.load('glue' , 'mrpc' ) set_seed(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = get_dataloaders(_UpperCAmelCase , _UpperCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase__ : Optional[int] = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=_UpperCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase__ : Tuple = model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase__ : Optional[int] = AdamW(params=model.parameters() , lr=_UpperCAmelCase ) # Instantiate scheduler lowerCamelCase__ : int = get_linear_schedule_with_warmup( optimizer=_UpperCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_UpperCAmelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = accelerator.prepare( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Now we train the model for epoch in range(_UpperCAmelCase ): model.train() for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_UpperCAmelCase ): lowerCamelCase__ : Any = model(**_UpperCAmelCase ) lowerCamelCase__ : str = output.loss accelerator.backward(_UpperCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase__ : Optional[Any] = model(**_UpperCAmelCase ) lowerCamelCase__ : List[Any] = outputs.logits.argmax(dim=-1 ) lowerCamelCase__ , lowerCamelCase__ : List[Any] = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=_UpperCAmelCase , references=_UpperCAmelCase , ) lowerCamelCase__ : List[str] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : Optional[int] = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=_UpperCAmelCase , default=_UpperCAmelCase , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=_UpperCAmelCase , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) lowerCamelCase__ : Optional[Any] = parser.parse_args() lowerCamelCase__ : Tuple = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
50
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """M-CLIP""" def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=1024 , UpperCAmelCase : Tuple=768 , **UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[int] = transformerDimSize lowerCamelCase__ : Optional[Any] = imageDimSize super().__init__(**UpperCAmelCase ) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = MCLIPConfig def __init__( self : List[Any] , UpperCAmelCase : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Dict: super().__init__(UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Tuple = XLMRobertaModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict ) -> Tuple: lowerCamelCase__ : Any = self.transformer(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase )[0] lowerCamelCase__ : int = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(UpperCAmelCase ), embs
50
1
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: return {key.lstrip('-' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : Optional[Any] = ArgumentParser( 'HuggingFace Datasets CLI tool' , usage='datasets-cli <command> [<args>]' , allow_abbrev=_UpperCAmelCase ) lowerCamelCase__ : List[Any] = parser.add_subparsers(help='datasets-cli command helpers' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(_UpperCAmelCase ) EnvironmentCommand.register_subcommand(_UpperCAmelCase ) TestCommand.register_subcommand(_UpperCAmelCase ) RunBeamCommand.register_subcommand(_UpperCAmelCase ) DummyDataCommand.register_subcommand(_UpperCAmelCase ) # Parse args lowerCamelCase__ , lowerCamelCase__ : Optional[int] = parser.parse_known_args() if not hasattr(_UpperCAmelCase , 'func' ): parser.print_help() exit(1 ) lowerCamelCase__ : Optional[Any] = parse_unknown_args(_UpperCAmelCase ) # Run lowerCamelCase__ : Optional[int] = args.func(_UpperCAmelCase , **_UpperCAmelCase ) service.run() if __name__ == "__main__": main()
50
from itertools import count def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Optional[Any] = [1] * min_block_length for n in count(_UpperCAmelCase ): fill_count_functions.append(1 ) for block_length in range(_UpperCAmelCase , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from timeit import timeit _UpperCAmelCase : Union[str, Any] = { """MALAYALAM""": True, """String""": False, """rotor""": True, """level""": True, """A""": True, """BB""": True, """ABC""": False, """amanaplanacanalpanama""": True, # "a man a plan a canal panama" } # Ensure our test data is valid assert all((key == key[::-1]) is value for key, value in test_data.items()) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: lowerCamelCase__ : Tuple = 0 lowerCamelCase__ : Dict = len(_UpperCAmelCase ) - 1 while start_i < end_i: if s[start_i] == s[end_i]: start_i += 1 end_i -= 1 else: return False return True def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: lowerCamelCase__ : Dict = len(_UpperCAmelCase ) // 2 lowerCamelCase__ : Tuple = len(_UpperCAmelCase ) # We need to traverse till half of the length of string # as we can get access of the i'th last element from # i'th index. # eg: [0,1,2,3,4,5] => 4th index can be accessed # with the help of 1st index (i==n-i-1) # where n is length of string return all(s[i] == s[n - i - 1] for i in range(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if len(_UpperCAmelCase ) <= 2: return True if s[0] == s[len(_UpperCAmelCase ) - 1]: return is_palindrome_recursive(s[1:-1] ) else: return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return s == s[::-1] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: lowerCamelCase__ : Any = F"""all({name}(key) is value for key, value in test_data.items())""" lowerCamelCase__ : List[Any] = F"""from __main__ import test_data, {name}""" lowerCamelCase__ : int = 50_0000 lowerCamelCase__ : List[str] = timeit(stmt=_UpperCAmelCase , setup=_UpperCAmelCase , number=_UpperCAmelCase ) print(F"""{name:<35} finished {number:,} runs in {result:.5f} seconds""" ) if __name__ == "__main__": for key, value in test_data.items(): assert is_palindrome(key) is is_palindrome_recursive(key) assert is_palindrome(key) is is_palindrome_slice(key) print(F"""{key:21} {value}""") print("""a man a plan a canal panama""") # finished 500,000 runs in 0.46793 seconds benchmark_function("""is_palindrome_slice""") # finished 500,000 runs in 0.85234 seconds benchmark_function("""is_palindrome""") # finished 500,000 runs in 1.32028 seconds benchmark_function("""is_palindrome_recursive""") # finished 500,000 runs in 2.08679 seconds benchmark_function("""is_palindrome_traversal""")
50
from __future__ import annotations from typing import Any def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: create_state_space_tree(_UpperCAmelCase , [] , 0 ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> None: if index == len(_UpperCAmelCase ): print(_UpperCAmelCase ) return create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _UpperCAmelCase : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["""A""", """B""", """C"""]) generate_all_subsequences(seq)
50
1
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : Tuple = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, oder?', } # BLUE scores as follows: # "pair": [fairseq, transformers] lowerCamelCase__ : Tuple = { 'ru-en': ['[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)', '39.20'], 'en-ru': ['[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)', '33.47'], 'en-de': ['[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)', '42.83'], 'de-en': ['[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)', '41.35'], } lowerCamelCase__ : str = F"""{src_lang}-{tgt_lang}""" lowerCamelCase__ : Any = F""" --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = \"facebook/wmt19-{src_lang}-{tgt_lang}\" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = \"{texts[src_lang]}\" input_ids = tokenizer.encode(input, return_tensors=\"pt\") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR's WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) """ os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , 'README.md' ) print(F"""Generating {path}""" ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(_UpperCAmelCase ) # make sure we are under the root of the project _UpperCAmelCase : Tuple = Path(__file__).resolve().parent.parent.parent _UpperCAmelCase : List[Any] = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase : int = model_name.split("""-""") _UpperCAmelCase : Tuple = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
50
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off _UpperCAmelCase : Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = NllbTokenizer UpperCAmelCase__ = [] UpperCAmelCase__ = [] def __init__( self : Tuple , UpperCAmelCase : int=None , UpperCAmelCase : Any=None , UpperCAmelCase : str="<s>" , UpperCAmelCase : Optional[Any]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Tuple="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[str]="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=False , **UpperCAmelCase : Optional[int] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token lowerCamelCase__ : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase , tokenizer_file=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , src_lang=UpperCAmelCase , tgt_lang=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , legacy_behaviour=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = vocab_file lowerCamelCase__ : Dict = False if not self.vocab_file else True lowerCamelCase__ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCamelCase__ : str = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : int = src_lang if src_lang is not None else 'eng_Latn' lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A_ ( self : int ) -> str: return self._src_lang @src_lang.setter def A_ ( self : List[Any] , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] , UpperCAmelCase : Optional[str] , **UpperCAmelCase : List[str] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCamelCase__ : Optional[int] = src_lang lowerCamelCase__ : Optional[int] = self(UpperCAmelCase , add_special_tokens=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.convert_tokens_to_ids(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tgt_lang_id return inputs def A_ ( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : str = "eng_Latn" , UpperCAmelCase : Optional[List[str]] = None , UpperCAmelCase : str = "fra_Latn" , **UpperCAmelCase : Dict , ) -> BatchEncoding: lowerCamelCase__ : Any = src_lang lowerCamelCase__ : int = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Optional[int]: return self.set_src_lang_special_tokens(self.src_lang ) def A_ ( self : Any ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A_ ( self : str , UpperCAmelCase : Optional[Any] ) -> None: lowerCamelCase__ : int = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : int = [self.cur_lang_code] lowerCamelCase__ : Tuple = [self.eos_token_id] lowerCamelCase__ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : int , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Union[str, Any] = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : Dict = [] lowerCamelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : Any = [self.cur_lang_code] lowerCamelCase__ : Optional[Any] = [self.eos_token_id] lowerCamelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCamelCase__ : int = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
1
from datetime import datetime as dt import os from github import Github _UpperCAmelCase : Dict = [ """good first issue""", """good second issue""", """good difficult issue""", """feature request""", """new model""", """wip""", ] def SCREAMING_SNAKE_CASE ( ) -> int: lowerCamelCase__ : Optional[Any] = Github(os.environ['GITHUB_TOKEN'] ) lowerCamelCase__ : Union[str, Any] = g.get_repo('huggingface/transformers' ) lowerCamelCase__ : int = repo.get_issues(state='open' ) for issue in open_issues: lowerCamelCase__ : str = sorted([comment for comment in issue.get_comments()] , key=lambda _UpperCAmelCase : i.created_at , reverse=_UpperCAmelCase ) lowerCamelCase__ : int = comments[0] if len(_UpperCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='closed' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
50
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ , lowerCamelCase__ : List[str] = emb.weight.shape lowerCamelCase__ : Tuple = nn.Linear(_UpperCAmelCase , _UpperCAmelCase , bias=_UpperCAmelCase ) lowerCamelCase__ : Dict = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location='cpu' ) lowerCamelCase__ : List[str] = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCamelCase__ : Optional[int] = mam_aaa['model'] remove_ignore_keys_(_UpperCAmelCase ) lowerCamelCase__ : str = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCamelCase__ : Union[str, Any] = MaMaaaConfig( vocab_size=_UpperCAmelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCamelCase__ : Optional[Any] = state_dict['decoder.embed_tokens.weight'] lowerCamelCase__ : Union[str, Any] = MaMaaaForConditionalGeneration(_UpperCAmelCase ) model.model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) lowerCamelCase__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase : str = parser.parse_args() _UpperCAmelCase : Optional[Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Union[str, Any] = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
50
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
1
import socket def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : Dict = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) lowerCamelCase__ : Union[str, Any] = socket.gethostname() lowerCamelCase__ : int = 1_2312 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: lowerCamelCase__ : Dict = sock.recv(1024 ) if not data: break out_file.write(_UpperCAmelCase ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
50
import flax.linen as nn import jax import jax.numpy as jnp class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , UpperCAmelCase : Dict ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = hidden_states.shape lowerCamelCase__ : Union[str, Any] = jax.image.resize( UpperCAmelCase , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : str , UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = None UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels lowerCamelCase__ : Tuple = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : int = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Union[str, Any] = nn.Dense(UpperCAmelCase , dtype=self.dtype ) lowerCamelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : List[Any] = nn.Dropout(self.dropout_prob ) lowerCamelCase__ : Tuple = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Optional[Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCamelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCamelCase__ : Dict = nn.Conv( UpperCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=True ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = hidden_states lowerCamelCase__ : List[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[Any] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Any = self.conva(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.time_emb_proj(nn.swish(UpperCAmelCase ) ) lowerCamelCase__ : List[str] = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase , 1 ) , 1 ) lowerCamelCase__ : List[str] = hidden_states + temb lowerCamelCase__ : Optional[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[str] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.dropout(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : str = self.conva(UpperCAmelCase ) if self.conv_shortcut is not None: lowerCamelCase__ : Dict = self.conv_shortcut(UpperCAmelCase ) return hidden_states + residual
50
1
import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=1024 ) -> int: lowerCamelCase__ , lowerCamelCase__ : List[str] = [], [] lowerCamelCase__ : Any = list(zip(_UpperCAmelCase , _UpperCAmelCase ) ) lowerCamelCase__ , lowerCamelCase__ : Tuple = sorted_examples[0] def is_too_big(_UpperCAmelCase ): return tok(_UpperCAmelCase , return_tensors='pt' ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): lowerCamelCase__ : Dict = new_src + ' ' + src lowerCamelCase__ : Any = new_tgt + ' ' + tgt if is_too_big(_UpperCAmelCase ) or is_too_big(_UpperCAmelCase ): # cant fit, finalize example finished_src.append(_UpperCAmelCase ) finished_tgt.append(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Tuple = src, tgt else: # can fit, keep adding lowerCamelCase__ , lowerCamelCase__ : Any = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(_UpperCAmelCase ) finished_tgt.append(_UpperCAmelCase ) return finished_src, finished_tgt def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : str = Path(_UpperCAmelCase ) save_path.mkdir(exist_ok=_UpperCAmelCase ) for split in ["train"]: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" lowerCamelCase__ : Optional[Any] = [x.rstrip() for x in Path(_UpperCAmelCase ).open().readlines()] lowerCamelCase__ : int = [x.rstrip() for x in Path(_UpperCAmelCase ).open().readlines()] lowerCamelCase__ , lowerCamelCase__ : Dict = pack_examples(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) print(F"""packed {split} split from {len(_UpperCAmelCase )} examples -> {len(_UpperCAmelCase )}.""" ) Path(save_path / F"""{split}.source""" ).open('w' ).write('\n'.join(_UpperCAmelCase ) ) Path(save_path / F"""{split}.target""" ).open('w' ).write('\n'.join(_UpperCAmelCase ) ) for split in ["val", "test"]: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" shutil.copyfile(_UpperCAmelCase , save_path / F"""{split}.source""" ) shutil.copyfile(_UpperCAmelCase , save_path / F"""{split}.target""" ) def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : Tuple = argparse.ArgumentParser() parser.add_argument('--tok_name' , type=_UpperCAmelCase , help='like facebook/bart-large-cnn,t5-base, etc.' ) parser.add_argument('--max_seq_len' , type=_UpperCAmelCase , default=128 ) parser.add_argument('--data_dir' , type=_UpperCAmelCase ) parser.add_argument('--save_path' , type=_UpperCAmelCase ) lowerCamelCase__ : List[str] = parser.parse_args() lowerCamelCase__ : List[Any] = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(_UpperCAmelCase , Path(args.data_dir ) , args.max_seq_len , args.save_path ) if __name__ == "__main__": packer_cli()
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Optional[Any] = set() # edges = list of graph's edges lowerCamelCase__ : List[str] = get_edges(_UpperCAmelCase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCamelCase__ , lowerCamelCase__ : str = edges.pop() chosen_vertices.add(_UpperCAmelCase ) chosen_vertices.add(_UpperCAmelCase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_UpperCAmelCase ) return chosen_vertices def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Union[str, Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
50
1
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: if depth < 0: raise ValueError('Depth cannot be less than 0' ) if not scores: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , ) ) def SCREAMING_SNAKE_CASE ( ) -> None: lowerCamelCase__ : int = [90, 23, 6, 33, 21, 65, 123, 3_4423] lowerCamelCase__ : Any = math.log(len(_UpperCAmelCase ) , 2 ) print(F"""Optimal value : {minimax(0 , 0 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
50
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
1
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Any = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """bart""" UpperCAmelCase__ = ["""past_key_values"""] UpperCAmelCase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Tuple , UpperCAmelCase : Optional[int]=50265 , UpperCAmelCase : Optional[int]=1024 , UpperCAmelCase : Optional[int]=12 , UpperCAmelCase : Dict=4096 , UpperCAmelCase : List[str]=16 , UpperCAmelCase : List[Any]=12 , UpperCAmelCase : List[Any]=4096 , UpperCAmelCase : Optional[Any]=16 , UpperCAmelCase : int=0.0 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : Optional[int]="gelu" , UpperCAmelCase : Tuple=1024 , UpperCAmelCase : int=0.1 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : str=0.0 , UpperCAmelCase : Tuple=0.0_2 , UpperCAmelCase : List[str]=0.0 , UpperCAmelCase : Any=False , UpperCAmelCase : List[str]=True , UpperCAmelCase : str=3 , UpperCAmelCase : Union[str, Any]=1 , UpperCAmelCase : List[str]=0 , UpperCAmelCase : List[Any]=2 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=2 , UpperCAmelCase : List[str]=2 , **UpperCAmelCase : Dict , ) -> Union[str, Any]: lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Tuple = max_position_embeddings lowerCamelCase__ : Optional[Any] = d_model lowerCamelCase__ : int = encoder_ffn_dim lowerCamelCase__ : int = encoder_layers lowerCamelCase__ : Optional[Any] = encoder_attention_heads lowerCamelCase__ : Dict = decoder_ffn_dim lowerCamelCase__ : Union[str, Any] = decoder_layers lowerCamelCase__ : List[Any] = decoder_attention_heads lowerCamelCase__ : List[str] = dropout lowerCamelCase__ : List[Any] = attention_dropout lowerCamelCase__ : Optional[Any] = activation_dropout lowerCamelCase__ : str = activation_function lowerCamelCase__ : List[str] = init_std lowerCamelCase__ : Optional[Any] = encoder_layerdrop lowerCamelCase__ : Optional[int] = decoder_layerdrop lowerCamelCase__ : int = classifier_dropout lowerCamelCase__ : Union[str, Any] = use_cache lowerCamelCase__ : str = encoder_layers lowerCamelCase__ : str = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=UpperCAmelCase , pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , is_encoder_decoder=UpperCAmelCase , decoder_start_token_id=UpperCAmelCase , forced_eos_token_id=UpperCAmelCase , **UpperCAmelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , UpperCAmelCase ): lowerCamelCase__ : Dict = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ 'The config can simply be saved and uploaded again to be fixed.' ) class lowerCAmelCase ( __UpperCamelCase ): @property def A_ ( self : Dict ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Any = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ : Tuple = {0: 'batch'} lowerCamelCase__ : Tuple = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: lowerCamelCase__ : Optional[Any] = {0: 'batch', 1: 'decoder_sequence'} lowerCamelCase__ : Dict = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(UpperCAmelCase , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. lowerCamelCase__ : int = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.num_layers for i in range(UpperCAmelCase ): lowerCamelCase__ : List[Any] = {0: 'batch', 2: 'past_sequence + sequence'} lowerCamelCase__ : Optional[Any] = {0: 'batch', 2: 'past_sequence + sequence'} else: lowerCamelCase__ : Union[str, Any] = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def A_ ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : int = super().outputs else: lowerCamelCase__ : Optional[Any] = super(UpperCAmelCase , self ).outputs if self.use_past: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.num_layers for i in range(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = {0: 'batch', 2: 'past_sequence + sequence'} lowerCamelCase__ : List[str] = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def A_ ( self : Optional[Any] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowerCamelCase__ : List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # Generate decoder inputs lowerCamelCase__ : List[str] = seq_length if not self.use_past else 1 lowerCamelCase__ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : List[str] = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} lowerCamelCase__ : Optional[int] = dict(**UpperCAmelCase , **UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ , lowerCamelCase__ : Optional[int] = common_inputs['input_ids'].shape lowerCamelCase__ : List[str] = common_inputs['decoder_input_ids'].shape[1] lowerCamelCase__ , lowerCamelCase__ : List[str] = self.num_attention_heads lowerCamelCase__ : Any = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Dict = decoder_seq_length + 3 lowerCamelCase__ : int = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowerCamelCase__ : Dict = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(UpperCAmelCase , UpperCAmelCase )] , dim=1 ) lowerCamelCase__ : Any = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowerCamelCase__ , lowerCamelCase__ : Tuple = self.num_layers lowerCamelCase__ : List[Any] = min(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Tuple = max(UpperCAmelCase , UpperCAmelCase ) - min_num_layers lowerCamelCase__ : str = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(UpperCAmelCase ): common_inputs["past_key_values"].append( ( torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase ), ) ) # TODO: test this. lowerCamelCase__ : List[str] = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(UpperCAmelCase , UpperCAmelCase ): common_inputs["past_key_values"].append((torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase )) ) return common_inputs def A_ ( self : Tuple , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowerCamelCase__ : Any = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowerCamelCase__ , lowerCamelCase__ : int = common_inputs['input_ids'].shape # Not using the same length for past_key_values lowerCamelCase__ : Union[str, Any] = seqlen + 2 lowerCamelCase__ , lowerCamelCase__ : List[Any] = self.num_layers lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = self.num_attention_heads lowerCamelCase__ : Tuple = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowerCamelCase__ : Union[str, Any] = common_inputs['attention_mask'].dtype lowerCamelCase__ : int = torch.cat( [common_inputs['attention_mask'], torch.ones(UpperCAmelCase , UpperCAmelCase , dtype=UpperCAmelCase )] , dim=1 ) lowerCamelCase__ : Any = [ (torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase )) for _ in range(UpperCAmelCase ) ] return common_inputs def A_ ( self : Optional[int] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowerCamelCase__ : Tuple = compute_effective_axis_dimension( UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowerCamelCase__ : List[str] = tokenizer.num_special_tokens_to_add(UpperCAmelCase ) lowerCamelCase__ : List[str] = compute_effective_axis_dimension( UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=UpperCAmelCase ) # Generate dummy inputs according to compute batch and sequence lowerCamelCase__ : str = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size lowerCamelCase__ : Dict = dict(tokenizer(UpperCAmelCase , return_tensors=UpperCAmelCase ) ) return common_inputs def A_ ( self : List[Any] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : List[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) elif self.task == "causal-lm": lowerCamelCase__ : int = self._generate_dummy_inputs_for_causal_lm( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) else: lowerCamelCase__ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) return common_inputs def A_ ( self : List[str] , UpperCAmelCase : Tuple , UpperCAmelCase : Any , UpperCAmelCase : Any , UpperCAmelCase : int ) -> List[Any]: if self.task in ["default", "seq2seq-lm"]: lowerCamelCase__ : Dict = super()._flatten_past_key_values_(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) else: lowerCamelCase__ : str = super(UpperCAmelCase , self )._flatten_past_key_values_( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase )
50
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
1
import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( SeqaSeqDataset, calculate_bleu, calculate_rouge, chunks, lmap, load_json, parse_numeric_n_bool_cl_kwargs, save_json, use_task_specific_params, write_txt_file, ) _UpperCAmelCase : Union[str, Any] = getLogger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 8 , _UpperCAmelCase = 1024 , _UpperCAmelCase="val" , _UpperCAmelCase=None , _UpperCAmelCase=False , _UpperCAmelCase="summarization" , _UpperCAmelCase=None , _UpperCAmelCase=1 , _UpperCAmelCase = None , _UpperCAmelCase="" , **_UpperCAmelCase , ) -> Dict: lowerCamelCase__ : List[str] = str(_UpperCAmelCase ) assert local_rank is not None torch.distributed.init_process_group(backend='nccl' , rank=_UpperCAmelCase ) lowerCamelCase__ : str = Path(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = save_dir.joinpath(F"""rank_{local_rank}_output.json""" ) torch.cuda.set_device(_UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoModelForSeqaSeqLM.from_pretrained(_UpperCAmelCase ).cuda() if fpaa: lowerCamelCase__ : List[str] = model.half() # determine if we need to increase num_beams use_task_specific_params(_UpperCAmelCase , _UpperCAmelCase ) # update config with task specific params lowerCamelCase__ : Optional[Any] = generate_kwargs.pop('num_beams' , model.config.num_beams ) # AttributeError risk? if num_return_sequences > num_beams: lowerCamelCase__ : List[str] = num_return_sequences lowerCamelCase__ : List[Any] = AutoTokenizer.from_pretrained(_UpperCAmelCase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. if max_source_length is None: lowerCamelCase__ : Tuple = tokenizer.model_max_length if prefix is None: lowerCamelCase__ : List[str] = prefix or getattr(model.config , 'prefix' , '' ) or '' lowerCamelCase__ : List[str] = SeqaSeqDataset( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , max_target_length=1024 , type_path=_UpperCAmelCase , n_obs=_UpperCAmelCase , prefix=_UpperCAmelCase , **_UpperCAmelCase , ) # I set shuffle=True for a more accurate progress bar. # If all the longest samples are first, the prog bar estimate is too high at the beginning. lowerCamelCase__ : Optional[Any] = ds.make_sortish_sampler(_UpperCAmelCase , distributed=_UpperCAmelCase , add_extra_examples=_UpperCAmelCase , shuffle=_UpperCAmelCase ) lowerCamelCase__ : str = DataLoader(_UpperCAmelCase , sampler=_UpperCAmelCase , batch_size=_UpperCAmelCase , collate_fn=ds.collate_fn ) lowerCamelCase__ : Dict = [] for batch in tqdm(_UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = model.generate( input_ids=batch['input_ids'].to(model.device ) , attention_mask=batch['attention_mask'].to(model.device ) , num_return_sequences=_UpperCAmelCase , num_beams=_UpperCAmelCase , **_UpperCAmelCase , ) lowerCamelCase__ : Dict = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase ) lowerCamelCase__ : Any = batch['ids'] if num_return_sequences > 1: lowerCamelCase__ : int = chunks(_UpperCAmelCase , _UpperCAmelCase ) # batch size chunks, each of size num_return_seq for i, pred in enumerate(_UpperCAmelCase ): results.append({'pred': pred, 'id': ids[i].item()} ) save_json(_UpperCAmelCase , _UpperCAmelCase ) return results, sampler.num_replicas def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : Dict = argparse.ArgumentParser( epilog='Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate' ) parser.add_argument('--data_dir' , type=_UpperCAmelCase , help='like cnn_dm/test.source' ) parser.add_argument( '--model_name' , type=_UpperCAmelCase , help='like facebook/bart-large-cnn,t5-base, etc.' , default='sshleifer/distilbart-xsum-12-3' , ) parser.add_argument('--save_dir' , type=_UpperCAmelCase , help='where to save' , default='tmp_gen' ) parser.add_argument('--max_source_length' , type=_UpperCAmelCase , default=_UpperCAmelCase ) parser.add_argument( '--type_path' , type=_UpperCAmelCase , default='test' , help='which subset to evaluate typically train/val/test' ) parser.add_argument('--task' , type=_UpperCAmelCase , default='summarization' , help='used for task_specific_params + metrics' ) parser.add_argument('--bs' , type=_UpperCAmelCase , default=8 , required=_UpperCAmelCase , help='batch size' ) parser.add_argument( '--local_rank' , type=_UpperCAmelCase , default=-1 , required=_UpperCAmelCase , help='should be passed by distributed.launch' ) parser.add_argument( '--n_obs' , type=_UpperCAmelCase , default=_UpperCAmelCase , required=_UpperCAmelCase , help='How many observations. Defaults to all.' ) parser.add_argument( '--num_return_sequences' , type=_UpperCAmelCase , default=1 , required=_UpperCAmelCase , help='How many sequences to return' ) parser.add_argument( '--sync_timeout' , type=_UpperCAmelCase , default=600 , required=_UpperCAmelCase , help='How long should master process wait for other processes to finish.' , ) parser.add_argument('--src_lang' , type=_UpperCAmelCase , default=_UpperCAmelCase , required=_UpperCAmelCase ) parser.add_argument('--tgt_lang' , type=_UpperCAmelCase , default=_UpperCAmelCase , required=_UpperCAmelCase ) parser.add_argument( '--prefix' , type=_UpperCAmelCase , required=_UpperCAmelCase , default=_UpperCAmelCase , help='will be added to the begininng of src examples' ) parser.add_argument('--fp16' , action='store_true' ) parser.add_argument('--debug' , action='store_true' ) lowerCamelCase__ : int = time.time() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = parser.parse_known_args() lowerCamelCase__ : Union[str, Any] = parse_numeric_n_bool_cl_kwargs(_UpperCAmelCase ) if generate_kwargs and args.local_rank <= 0: print(F"""parsed the following generate kwargs: {generate_kwargs}""" ) lowerCamelCase__ : Union[str, Any] = Path(args.save_dir + '_tmp' ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) # this handles locking. lowerCamelCase__ : str = list(json_save_dir.glob('rank_*.json' ) ) if intermediate_files: raise ValueError(F"""Found files at {json_save_dir} please move or remove them.""" ) # In theory, a node could finish and save before another node hits this. If this happens, we can address later. lowerCamelCase__ : Any = {} if args.src_lang is not None: lowerCamelCase__ : int = args.src_lang if args.tgt_lang is not None: lowerCamelCase__ : Union[str, Any] = args.tgt_lang Path(args.save_dir ).mkdir(exist_ok=_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Tuple = eval_data_dir( args.data_dir , _UpperCAmelCase , args.model_name , type_path=args.type_path , bs=args.bs , fpaa=args.fpaa , task=args.task , local_rank=args.local_rank , n_obs=args.n_obs , max_source_length=args.max_source_length , num_return_sequences=args.num_return_sequences , prefix=args.prefix , dataset_kwargs=_UpperCAmelCase , **_UpperCAmelCase , ) if args.local_rank <= 0: lowerCamelCase__ : Optional[Any] = Path(args.save_dir ) save_dir.mkdir(exist_ok=_UpperCAmelCase ) lowerCamelCase__ : List[str] = gather_results_from_each_node(_UpperCAmelCase , _UpperCAmelCase , args.sync_timeout ) lowerCamelCase__ : Dict = combine_partial_results(_UpperCAmelCase ) if args.num_return_sequences > 1: lowerCamelCase__ : Any = save_dir.joinpath('pseudolabel_results.json' ) print(F"""Saving aggregated results at {save_path}, intermediate in {json_save_dir}/""" ) save_json(_UpperCAmelCase , _UpperCAmelCase ) return lowerCamelCase__ : Optional[int] = Path(args.data_dir ).joinpath(args.type_path + '.target' ) with open(_UpperCAmelCase ) as f: lowerCamelCase__ : Optional[Any] = [x.rstrip() for x in f.readlines()][: len(_UpperCAmelCase )] # Calculate metrics, save metrics, and save _generations.txt lowerCamelCase__ : Tuple = 'translation' in args.task lowerCamelCase__ : List[str] = calculate_bleu if calc_bleu else calculate_rouge lowerCamelCase__ : List[Any] = 'bleu' if calc_bleu else 'rouge' lowerCamelCase__ : Dict = score_fn(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Optional[int] = len(_UpperCAmelCase ) lowerCamelCase__ : str = time.time() - start_time lowerCamelCase__ : int = round(runtime / metrics['n_obs'] , 4 ) lowerCamelCase__ : List[str] = num_replicas # TODO(@stas00): add whatever metadata to metrics lowerCamelCase__ : int = save_dir.joinpath(F"""{args.type_path}_{metric_name}.json""" ) save_json(_UpperCAmelCase , _UpperCAmelCase , indent=_UpperCAmelCase ) print(_UpperCAmelCase ) write_txt_file(_UpperCAmelCase , save_dir.joinpath(F"""{args.type_path}_generations.txt""" ) ) if args.debug: write_txt_file(_UpperCAmelCase , save_dir.joinpath(F"""{args.type_path}.target""" ) ) else: shutil.rmtree(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List: lowerCamelCase__ : Union[str, Any] = [] for partial_result in partial_results: records.extend(_UpperCAmelCase ) lowerCamelCase__ : Tuple = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x["id"] ) lowerCamelCase__ : Union[str, Any] = [x['pred'] for x in records] return preds def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[Dict[str, List]]: # WAIT FOR lots of .json files lowerCamelCase__ : Optional[Any] = time.time() logger.info('waiting for all nodes to finish' ) lowerCamelCase__ : List[str] = None while (time.time() - start_wait) < timeout: lowerCamelCase__ : Any = list(save_dir.glob('rank_*.json' ) ) if len(_UpperCAmelCase ) < num_replicas: continue try: # make sure all json files are fully saved lowerCamelCase__ : Any = lmap(_UpperCAmelCase , _UpperCAmelCase ) return json_data except JSONDecodeError: continue else: raise TimeoutError('Rank 0 gave up on waiting for other processes' ) # Unreachable if __name__ == "__main__": # Usage for MT: run_generate()
50
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : str=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : str=9 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=True , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : List[str]=5 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Union[str, Any]=37 , UpperCAmelCase : int=8 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=0.0_0_2 , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=0 , UpperCAmelCase : Union[str, Any]=0 , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , ) -> Union[str, Any]: lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = encoder_seq_length lowerCamelCase__ : int = decoder_seq_length # For common tests lowerCamelCase__ : List[str] = self.decoder_seq_length lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : List[Any] = use_attention_mask lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = d_ff lowerCamelCase__ : Optional[Any] = relative_attention_num_buckets lowerCamelCase__ : Any = dropout_rate lowerCamelCase__ : Any = initializer_factor lowerCamelCase__ : Union[str, Any] = eos_token_id lowerCamelCase__ : List[str] = pad_token_id lowerCamelCase__ : List[str] = decoder_start_token_id lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = decoder_layers def A_ ( self : List[Any] ) -> int: return TaConfig.from_pretrained('google/umt5-base' ) def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=None , ) -> List[str]: if attention_mask is None: lowerCamelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCamelCase__ : Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCamelCase__ : int = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase ) if decoder_head_mask is None: lowerCamelCase__ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A_ ( self : str ) -> List[str]: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCamelCase__ : List[str] = input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Union[str, Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Dict = self.get_config() lowerCamelCase__ : Tuple = config.num_attention_heads lowerCamelCase__ : Any = self.prepare_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, input_dict def A_ ( self : Tuple ) -> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Dict = self.prepare_config_and_inputs() return config, inputs_dict def A_ ( self : Optional[int] ) -> List[str]: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Union[str, Any] ) -> Dict: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Dict , ) -> str: lowerCamelCase__ : Dict = UMTaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , ) lowerCamelCase__ : Any = model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) lowerCamelCase__ : Dict = result.last_hidden_state lowerCamelCase__ : Any = result.past_key_values lowerCamelCase__ : List[Any] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[Any] , ) -> Optional[int]: lowerCamelCase__ : List[Any] = UMTaModel(config=UpperCAmelCase ).get_decoder().to(UpperCAmelCase ).eval() # first forward pass lowerCamelCase__ : Tuple = model(UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , use_cache=UpperCAmelCase ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) + 1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCamelCase__ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )['last_hidden_state'] lowerCamelCase__ : str = model(UpperCAmelCase , past_key_values=UpperCAmelCase )['last_hidden_state'] # select random slice lowerCamelCase__ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase__ : Tuple = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase__ : List[str] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase , UpperCAmelCase , atol=1e-3 ) ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = UMTaModel(config=UpperCAmelCase ).to(UpperCAmelCase ).half().eval() lowerCamelCase__ : Optional[int] = model(**UpperCAmelCase )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCAmelCase ).any().item() ) @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCAmelCase__ = (UMTaForConditionalGeneration,) if is_torch_available() else () UpperCAmelCase__ = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = True UpperCAmelCase__ = True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCAmelCase__ = [0.8, 0.9] def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Tuple = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCAmelCase , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: lowerCamelCase__ : int = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Any = config_and_inputs[0] lowerCamelCase__ : Any = UMTaForConditionalGeneration(UpperCAmelCase ).eval() model.to(UpperCAmelCase ) lowerCamelCase__ : Tuple = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), } for attn_name, (name, mask) in zip(UpperCAmelCase , head_masking.items() ): lowerCamelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCamelCase__ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ) lowerCamelCase__ : Tuple = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase , return_dict_in_generate=UpperCAmelCase , **UpperCAmelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCamelCase__ : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=UpperCAmelCase ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=UpperCAmelCase , legacy=UpperCAmelCase ) lowerCamelCase__ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowerCamelCase__ : Tuple = tokenizer(UpperCAmelCase , return_tensors='pt' , padding=UpperCAmelCase ).input_ids # fmt: off lowerCamelCase__ : Any = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model.generate(input_ids.to(UpperCAmelCase ) ) lowerCamelCase__ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowerCamelCase__ : Union[str, Any] = tokenizer.batch_decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
1
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np _UpperCAmelCase : Union[str, Any] = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 _UpperCAmelCase : int = typing.Union[np.floataa, int, float] # noqa: UP007 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> VectorOut: return np.sqrt(np.sum((np.asarray(_UpperCAmelCase ) - np.asarray(_UpperCAmelCase )) ** 2 ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> VectorOut: return sum((va - va) ** 2 for va, va in zip(_UpperCAmelCase , _UpperCAmelCase ) ) ** (1 / 2) if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( ) -> None: from timeit import timeit print('Without Numpy' ) print( timeit( 'euclidean_distance_no_np([1, 2, 3], [4, 5, 6])' , number=1_0000 , globals=globals() , ) ) print('With Numpy' ) print( timeit( 'euclidean_distance([1, 2, 3], [4, 5, 6])' , number=1_0000 , globals=globals() , ) ) benchmark()
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: _UpperCAmelCase : str = None _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : List[str] = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : Optional[int] = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } _UpperCAmelCase : Optional[Any] = { """albert-base-v1""": 5_12, """albert-large-v1""": 5_12, """albert-xlarge-v1""": 5_12, """albert-xxlarge-v1""": 5_12, """albert-base-v2""": 5_12, """albert-large-v2""": 5_12, """albert-xlarge-v2""": 5_12, """albert-xxlarge-v2""": 5_12, } _UpperCAmelCase : Optional[int] = """▁""" class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = AlbertTokenizer def __init__( self : Optional[Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : List[Any]=False , UpperCAmelCase : Tuple="[CLS]" , UpperCAmelCase : str="[SEP]" , UpperCAmelCase : Any="<unk>" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : List[Any]="<pad>" , UpperCAmelCase : List[str]="[CLS]" , UpperCAmelCase : Optional[int]="[MASK]" , **UpperCAmelCase : int , ) -> Dict: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowerCamelCase__ : Any = ( AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase , normalized=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token ) super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , remove_space=UpperCAmelCase , keep_accents=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : Union[str, Any] = do_lower_case lowerCamelCase__ : List[Any] = remove_space lowerCamelCase__ : Optional[Any] = keep_accents lowerCamelCase__ : Tuple = vocab_file lowerCamelCase__ : Tuple = False if not self.vocab_file else True def A_ ( self : List[str] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Optional[Any] = [self.sep_token_id] lowerCamelCase__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A_ ( self : List[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[str] = [self.sep_token_id] lowerCamelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Any , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : List[str] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase ): @register_to_config def __init__( self : List[str] , UpperCAmelCase : int = 65536 , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 0 , UpperCAmelCase : str = "fourier" , UpperCAmelCase : bool = True , UpperCAmelCase : bool = False , UpperCAmelCase : float = 0.0 , UpperCAmelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase : str = None , UpperCAmelCase : Tuple[int] = (32, 32, 64) , UpperCAmelCase : str = None , UpperCAmelCase : int = 8 , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = False , ) -> List[Any]: super().__init__() lowerCamelCase__ : Optional[int] = sample_size # time if time_embedding_type == "fourier": lowerCamelCase__ : Optional[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase , log=UpperCAmelCase , flip_sin_to_cos=UpperCAmelCase ) lowerCamelCase__ : Any = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCamelCase__ : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase , downscale_freq_shift=UpperCAmelCase ) lowerCamelCase__ : Dict = block_out_channels[0] if use_timestep_embedding: lowerCamelCase__ : str = block_out_channels[0] * 4 lowerCamelCase__ : List[Any] = TimestepEmbedding( in_channels=UpperCAmelCase , time_embed_dim=UpperCAmelCase , act_fn=UpperCAmelCase , out_dim=block_out_channels[0] , ) lowerCamelCase__ : Any = nn.ModuleList([] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Optional[int] = None # down lowerCamelCase__ : Optional[int] = in_channels for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = output_channel lowerCamelCase__ : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Optional[int] = get_down_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Optional[int] = get_mid_block( UpperCAmelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase , add_downsample=UpperCAmelCase , ) # up lowerCamelCase__ : Optional[int] = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = reversed_block_out_channels[0] if out_block_type is None: lowerCamelCase__ : List[str] = out_channels else: lowerCamelCase__ : Any = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Tuple = output_channel lowerCamelCase__ : Union[str, Any] = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase ) - 1 else final_upsample_channels ) lowerCamelCase__ : List[str] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Dict = get_up_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : int = output_channel # out lowerCamelCase__ : int = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCamelCase__ : List[Any] = get_out_block( out_block_type=UpperCAmelCase , num_groups_out=UpperCAmelCase , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase , act_fn=UpperCAmelCase , fc_dim=block_out_channels[-1] // 4 , ) def A_ ( self : List[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Union[torch.Tensor, float, int] , UpperCAmelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: lowerCamelCase__ : Optional[Any] = timestep if not torch.is_tensor(UpperCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(sample.device ) lowerCamelCase__ : Optional[int] = self.time_proj(UpperCAmelCase ) if self.config.use_timestep_embedding: lowerCamelCase__ : str = self.time_mlp(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = timestep_embed[..., None] lowerCamelCase__ : str = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCamelCase__ : str = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCamelCase__ : str = () for downsample_block in self.down_blocks: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = downsample_block(hidden_states=UpperCAmelCase , temb=UpperCAmelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCamelCase__ : Optional[Any] = self.mid_block(UpperCAmelCase , UpperCAmelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCamelCase__ : Dict = down_block_res_samples[-1:] lowerCamelCase__ : Optional[Any] = down_block_res_samples[:-1] lowerCamelCase__ : Any = upsample_block(UpperCAmelCase , res_hidden_states_tuple=UpperCAmelCase , temb=UpperCAmelCase ) # 5. post-process if self.out_block: lowerCamelCase__ : Any = self.out_block(UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase )
50
1
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """M-CLIP""" def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=1024 , UpperCAmelCase : Tuple=768 , **UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[int] = transformerDimSize lowerCamelCase__ : Optional[Any] = imageDimSize super().__init__(**UpperCAmelCase ) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = MCLIPConfig def __init__( self : List[Any] , UpperCAmelCase : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Dict: super().__init__(UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Tuple = XLMRobertaModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict ) -> Tuple: lowerCamelCase__ : Any = self.transformer(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase )[0] lowerCamelCase__ : int = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(UpperCAmelCase ), embs
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Dict: assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def SCREAMING_SNAKE_CASE ( ) -> List[Any]: assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : str = 'mock-s3-bucket' lowerCamelCase__ : Any = F"""s3://{mock_bucket}""" lowerCamelCase__ : int = extract_path_from_uri(_UpperCAmelCase ) assert dataset_path.startswith('s3://' ) is False lowerCamelCase__ : List[Any] = './local/path' lowerCamelCase__ : Tuple = extract_path_from_uri(_UpperCAmelCase ) assert dataset_path == new_dataset_path def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = is_remote_filesystem(_UpperCAmelCase ) assert is_remote is True lowerCamelCase__ : List[str] = fsspec.filesystem('file' ) lowerCamelCase__ : List[str] = is_remote_filesystem(_UpperCAmelCase ) assert is_remote is False @pytest.mark.parametrize('compression_fs_class' , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Any = {'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_file, 'bz2': bza_file, 'lz4': lza_file} lowerCamelCase__ : str = input_paths[compression_fs_class.protocol] if input_path is None: lowerCamelCase__ : List[str] = F"""for '{compression_fs_class.protocol}' compression protocol, """ if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_UpperCAmelCase ) lowerCamelCase__ : Dict = fsspec.filesystem(compression_fs_class.protocol , fo=_UpperCAmelCase ) assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Optional[int] = os.path.basename(_UpperCAmelCase ) lowerCamelCase__ : List[Any] = expected_filename[: expected_filename.rindex('.' )] assert fs.glob('*' ) == [expected_filename] with fs.open(_UpperCAmelCase , 'r' , encoding='utf-8' ) as f, open(_UpperCAmelCase , encoding='utf-8' ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize('protocol' , ['zip', 'gzip'] ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : Optional[Any] = {'zip': zip_jsonl_path, 'gzip': jsonl_gz_path} lowerCamelCase__ : List[str] = compressed_file_paths[protocol] lowerCamelCase__ : List[str] = 'dataset.jsonl' lowerCamelCase__ : Optional[Any] = F"""{protocol}://{member_file_path}::{compressed_file_path}""" lowerCamelCase__ , *lowerCamelCase__ : Optional[int] = fsspec.get_fs_token_paths(_UpperCAmelCase ) assert fs.isfile(_UpperCAmelCase ) assert not fs.isfile('non_existing_' + member_file_path ) @pytest.mark.integration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Optional[int] = hf_api.dataset_info(_UpperCAmelCase , token=_UpperCAmelCase ) lowerCamelCase__ : Any = HfFileSystem(repo_info=_UpperCAmelCase , token=_UpperCAmelCase ) assert sorted(hffs.glob('*' ) ) == [".gitattributes", "data"] assert hffs.isdir('data' ) assert hffs.isfile('.gitattributes' ) and hffs.isfile('data/text_data.txt' ) with open(_UpperCAmelCase ) as f: assert hffs.open('data/text_data.txt' , 'r' ).read() == f.read() def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : List[Any] = 'bz2' # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_UpperCAmelCase , _UpperCAmelCase , clobber=_UpperCAmelCase ) with pytest.warns(_UpperCAmelCase ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_UpperCAmelCase ) == 1 assert ( str(warning_info[0].message ) == F"""A filesystem protocol was already set for {protocol} and will be overwritten.""" )
50
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[int] = [] embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", F"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", F"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", F"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", F"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Tuple = [] attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", F"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", F"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", F"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", F"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] token.append((F"""cvt.encoder.stages.{idx}.cls_token""", 'stage2.cls_token') ) return token def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Tuple = 'imagenet-1k-id2label.json' lowerCamelCase__ : Union[str, Any] = 1000 lowerCamelCase__ : Optional[Any] = 'huggingface/label-files' lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowerCamelCase__ : int = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : List[str] = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowerCamelCase__ : List[Any] = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowerCamelCase__ : Dict = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowerCamelCase__ : Optional[Any] = [2, 2, 20] lowerCamelCase__ : Optional[int] = [3, 12, 16] lowerCamelCase__ : str = [192, 768, 1024] lowerCamelCase__ : Any = CvtForImageClassification(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowerCamelCase__ : Optional[int] = OrderedDict() lowerCamelCase__ : Tuple = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowerCamelCase__ : Optional[Any] = list_of_state_dict + cls_token(_UpperCAmelCase ) lowerCamelCase__ : str = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowerCamelCase__ : str = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : int = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : str = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( """--cvt_model""", default="""cvt-w24""", type=str, help="""Name of the cvt model you'd like to convert.""", ) parser.add_argument( """--image_size""", default=3_84, type=int, help="""Input Image Size""", ) parser.add_argument( """--cvt_file_name""", default=R"""cvtmodels\CvT-w24-384x384-IN-22k.pth""", type=str, help="""Input Image Size""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
50
1
import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = PriorTransformer UpperCAmelCase__ = """hidden_states""" @property def A_ ( self : List[str] ) -> str: lowerCamelCase__ : Union[str, Any] = 4 lowerCamelCase__ : List[str] = 8 lowerCamelCase__ : List[Any] = 7 lowerCamelCase__ : List[str] = floats_tensor((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : List[Any] = floats_tensor((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = floats_tensor((batch_size, num_embeddings, embedding_dim) ).to(UpperCAmelCase ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def A_ ( self : Union[str, Any] , UpperCAmelCase : Union[str, Any]=0 ) -> int: torch.manual_seed(UpperCAmelCase ) lowerCamelCase__ : int = 4 lowerCamelCase__ : Dict = 8 lowerCamelCase__ : Dict = 7 lowerCamelCase__ : List[Any] = torch.randn((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : str = torch.randn((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = torch.randn((batch_size, num_embeddings, embedding_dim) ).to(UpperCAmelCase ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def A_ ( self : Union[str, Any] ) -> List[str]: return (4, 8) @property def A_ ( self : List[str] ) -> Union[str, Any]: return (4, 8) def A_ ( self : Any ) -> int: lowerCamelCase__ : Tuple = { 'num_attention_heads': 2, 'attention_head_dim': 4, 'num_layers': 2, 'embedding_dim': 8, 'num_embeddings': 7, 'additional_embeddings': 4, } lowerCamelCase__ : List[str] = self.dummy_input return init_dict, inputs_dict def A_ ( self : List[Any] ) -> Tuple: lowerCamelCase__ , lowerCamelCase__ : List[str] = PriorTransformer.from_pretrained( 'hf-internal-testing/prior-dummy' , output_loading_info=UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) self.assertEqual(len(loading_info['missing_keys'] ) , 0 ) model.to(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model(**self.dummy_input )[0] assert hidden_states is not None, "Make sure output is not None" def A_ ( self : Tuple ) -> Tuple: lowerCamelCase__ , lowerCamelCase__ : str = self.prepare_init_args_and_inputs_for_common() lowerCamelCase__ : List[Any] = self.model_class(**UpperCAmelCase ) lowerCamelCase__ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Optional[int] = [*signature.parameters.keys()] lowerCamelCase__ : List[Any] = ['hidden_states', 'timestep'] self.assertListEqual(arg_names[:2] , UpperCAmelCase ) def A_ ( self : Optional[int] ) -> Optional[int]: lowerCamelCase__ : List[str] = PriorTransformer.from_pretrained('hf-internal-testing/prior-dummy' ) lowerCamelCase__ : int = model.to(UpperCAmelCase ) if hasattr(UpperCAmelCase , 'set_default_attn_processor' ): model.set_default_attn_processor() lowerCamelCase__ : Optional[int] = self.get_dummy_seed_input() with torch.no_grad(): lowerCamelCase__ : List[str] = model(**UpperCAmelCase )[0] lowerCamelCase__ : List[str] = output[0, :5].flatten().cpu() print(UpperCAmelCase ) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. lowerCamelCase__ : str = torch.tensor([-1.3_4_3_6, -0.2_8_7_0, 0.7_5_3_8, 0.4_3_6_8, -0.0_2_3_9] ) self.assertTrue(torch_all_close(UpperCAmelCase , UpperCAmelCase , rtol=1e-2 ) ) @slow class lowerCAmelCase ( unittest.TestCase ): def A_ ( self : str , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=768 , UpperCAmelCase : Tuple=77 , UpperCAmelCase : List[Any]=0 ) -> Optional[Any]: torch.manual_seed(UpperCAmelCase ) lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : List[str] = embedding_dim lowerCamelCase__ : Union[str, Any] = num_embeddings lowerCamelCase__ : Dict = torch.randn((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : Any = torch.randn((batch_size, embedding_dim) ).to(UpperCAmelCase ) lowerCamelCase__ : Dict = torch.randn((batch_size, num_embeddings, embedding_dim) ).to(UpperCAmelCase ) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def A_ ( self : Optional[int] ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [13, [-0.5_8_6_1, 0.1_2_8_3, -0.0_9_3_1, 0.0_8_8_2, 0.4_4_7_6, 0.1_3_2_9, -0.0_4_9_8, 0.0_6_4_0]], [37, [-0.4_9_1_3, 0.0_1_1_0, -0.0_4_8_3, 0.0_5_4_1, 0.4_9_5_4, -0.0_1_7_0, 0.0_3_5_4, 0.1_6_5_1]], # fmt: on ] ) def A_ ( self : List[str] , UpperCAmelCase : str , UpperCAmelCase : Tuple ) -> int: lowerCamelCase__ : Any = PriorTransformer.from_pretrained('kandinsky-community/kandinsky-2-1-prior' , subfolder='prior' ) model.to(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.get_dummy_seed_input(seed=UpperCAmelCase ) with torch.no_grad(): lowerCamelCase__ : List[str] = model(**UpperCAmelCase )[0] assert list(sample.shape ) == [1, 768] lowerCamelCase__ : List[str] = sample[0, :8].flatten().cpu() print(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = torch.tensor(UpperCAmelCase ) assert torch_all_close(UpperCAmelCase , UpperCAmelCase , atol=1e-3 )
50
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
1
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo _UpperCAmelCase : List[Any] = """\ @misc{wu2016googles, title={Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } """ _UpperCAmelCase : Union[str, Any] = """\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the 'GLEU score'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score's range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. """ _UpperCAmelCase : Optional[int] = """\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: 'google_bleu': google_bleu score Examples: Example 1: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.44 Example 2: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.61 Example 3: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results[\"google_bleu\"], 2)) 0.53 Example 4: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results[\"google_bleu\"], 2)) 0.4 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def A_ ( self : str ) -> MetricInfo: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ), 'references': datasets.Sequence( datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ) , id='references' ), } ) , ) def A_ ( self : Any , UpperCAmelCase : List[List[List[str]]] , UpperCAmelCase : List[List[str]] , UpperCAmelCase : int = 1 , UpperCAmelCase : int = 4 , ) -> Dict[str, float]: return { "google_bleu": gleu_score.corpus_gleu( list_of_references=UpperCAmelCase , hypotheses=UpperCAmelCase , min_len=UpperCAmelCase , max_len=UpperCAmelCase ) }
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase = " " ) -> list: lowerCamelCase__ : Tuple = [] lowerCamelCase__ : str = 0 for index, char in enumerate(_UpperCAmelCase ): if char == separator: split_words.append(string[last_index:index] ) lowerCamelCase__ : int = index + 1 elif index + 1 == len(_UpperCAmelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
50
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } _UpperCAmelCase : Any = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]="[UNK]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Tuple="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=None , **UpperCAmelCase : Optional[int] , ) -> str: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : str = strip_accents lowerCamelCase__ : Optional[Any] = tokenize_chinese_chars lowerCamelCase__ : int = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : str = do_lower_case def A_ ( self : Optional[int] , UpperCAmelCase : int , **UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH lowerCamelCase__ : Optional[int] = text lowerCamelCase__ : Dict = kwargs.pop('text_pair' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = kwargs.pop('return_tensors' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: lowerCamelCase__ : Tuple = batch_text_pair[idx] else: lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Any = encoded_candidates.get('input_ids' ) lowerCamelCase__ : Union[str, Any] = encoded_candidates.get('attention_mask' ) lowerCamelCase__ : Tuple = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) lowerCamelCase__ : int = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ) -> List[str]: lowerCamelCase__ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
1
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Dict = checkpoints.load_tax_checkpoint(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = flatten_dict(_UpperCAmelCase ) return flax_params def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[Any]: lowerCamelCase__ : Union[str, Any] = {} lowerCamelCase__ : List[str] = { 'token_embedder': 'embeddings', 'encoder_norm': 'layernorm', 'kernel': 'weight', '.out': '.output', 'scale': 'weight', 'embedders_0.pos_embedding': 'row_embedder.weight', 'embedders_1.pos_embedding': 'column_embedder.weight', } lowerCamelCase__ : int = { 'query': 'attention.query', 'key': 'attention.key', 'value': 'attention.value', 'output.dense': 'output', 'encoder_decoder_attention.o': 'encoder_decoder_attention.attention.o', 'pre_self_attention_layer_norm': 'self_attention.layer_norm', 'pre_cross_attention_layer_norm': 'encoder_decoder_attention.layer_norm', 'mlp.': 'mlp.DenseReluDense.', 'pre_mlp_layer_norm': 'mlp.layer_norm', 'self_attention.o': 'self_attention.attention.o', 'decoder.embeddings.embedding': 'decoder.embed_tokens.weight', 'decoder.relpos_bias.rel_embedding': 'decoder.layer.0.self_attention.attention.relative_attention_bias.weight', 'decoder.decoder_norm.weight': 'decoder.final_layer_norm.weight', 'decoder.logits_dense.weight': 'decoder.lm_head.weight', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowerCamelCase__ : Union[str, Any] = '.'.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowerCamelCase__ : str = new_key.replace(_UpperCAmelCase , _UpperCAmelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowerCamelCase__ : int = new_key.replace(_UpperCAmelCase , _UpperCAmelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowerCamelCase__ : Tuple = re.sub(r'layers_(\d+)' , r'layer.\1' , _UpperCAmelCase ) lowerCamelCase__ : Dict = new_key.replace('encoder' , 'encoder.encoder' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowerCamelCase__ : str = re.sub(r'layers_(\d+)' , r'layer.\1' , _UpperCAmelCase ) lowerCamelCase__ : Optional[int] = flax_dict[key] lowerCamelCase__ : Any = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowerCamelCase__ : Tuple = torch.from_numpy(converted_dict[key].T ) else: lowerCamelCase__ : Union[str, Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=False ) -> List[str]: lowerCamelCase__ : Any = get_flax_param(_UpperCAmelCase ) if not use_large: lowerCamelCase__ : str = PixaStructVisionConfig() lowerCamelCase__ : Dict = PixaStructTextConfig() else: lowerCamelCase__ : Dict = PixaStructVisionConfig( hidden_size=1536 , d_ff=3968 , num_attention_heads=24 , num_hidden_layers=18 ) lowerCamelCase__ : Optional[Any] = PixaStructTextConfig(hidden_size=1536 , d_ff=3968 , num_heads=24 , num_layers=18 ) lowerCamelCase__ : Any = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = PixaStructForConditionalGeneration(_UpperCAmelCase ) lowerCamelCase__ : Tuple = rename_and_convert_flax_params(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) lowerCamelCase__ : str = AutoTokenizer.from_pretrained('ybelkada/test-pix2struct-tokenizer' ) lowerCamelCase__ : List[str] = PixaStructImageProcessor() lowerCamelCase__ : Dict = PixaStructProcessor(image_processor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) if use_large: lowerCamelCase__ : Any = 4096 lowerCamelCase__ : Tuple = True # mkdir if needed os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) print('Model saved in {}'.format(_UpperCAmelCase ) ) if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--t5x_checkpoint_path""", default=None, type=str, help="""Path to the original T5x checkpoint.""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--use_large""", action="""store_true""", help="""Use large model.""") parser.add_argument("""--is_vqa""", action="""store_true""", help="""Use large model.""") _UpperCAmelCase : Optional[int] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
50
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _UpperCAmelCase : Optional[Any] = logging.getLogger(__name__) @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if is_torch_available(): import torch from torch.utils.data import Dataset class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 def __init__( self : int , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : List[str]=False , UpperCAmelCase : bool = False , ) -> List[str]: lowerCamelCase__ : int = hans_processors[task]() lowerCamelCase__ : Optional[Any] = os.path.join( UpperCAmelCase , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase ) , UpperCAmelCase , ) , ) lowerCamelCase__ : int = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = label_list[2], label_list[1] lowerCamelCase__ : List[str] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : str = cached_features_file + '.lock' with FileLock(UpperCAmelCase ): if os.path.exists(UpperCAmelCase ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) lowerCamelCase__ : int = torch.load(UpperCAmelCase ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) lowerCamelCase__ : str = ( processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) ) logger.info('Training examples: %s' , len(UpperCAmelCase ) ) lowerCamelCase__ : Dict = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) logger.info('Saving features into cached file %s' , UpperCAmelCase ) torch.save(self.features , UpperCAmelCase ) def __len__( self : Optional[int] ) -> Optional[Any]: return len(self.features ) def __getitem__( self : Tuple , UpperCAmelCase : Dict ) -> InputFeatures: return self.features[i] def A_ ( self : int ) -> int: return self.label_list if is_tf_available(): import tensorflow as tf class lowerCAmelCase : UpperCAmelCase__ = 42 def __init__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = 128 , UpperCAmelCase : Any=False , UpperCAmelCase : bool = False , ) -> Union[str, Any]: lowerCamelCase__ : Any = hans_processors[task]() lowerCamelCase__ : Optional[Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : str = label_list[2], label_list[1] lowerCamelCase__ : Optional[int] = label_list lowerCamelCase__ : int = processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCamelCase__ : Optional[int] = tf.data.Dataset.from_generator( UpperCAmelCase , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def A_ ( self : Any ) -> Any: return self.dataset def __len__( self : Tuple ) -> int: return len(self.features ) def __getitem__( self : List[str] , UpperCAmelCase : Any ) -> InputFeatures: return self.features[i] def A_ ( self : Dict ) -> str: return self.label_list class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : int , UpperCAmelCase : List[Any] ) -> int: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_train_set.txt' ) ) , 'train' ) def A_ ( self : Any , UpperCAmelCase : int ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def A_ ( self : Any ) -> List[Any]: return ["contradiction", "entailment", "neutral"] def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> List[str]: lowerCamelCase__ : List[str] = [] for i, line in enumerate(UpperCAmelCase ): if i == 0: continue lowerCamelCase__ : Tuple = '%s-%s' % (set_type, line[0]) lowerCamelCase__ : str = line[5] lowerCamelCase__ : Dict = line[6] lowerCamelCase__ : int = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCamelCase__ : Dict = line[0] examples.append(InputExample(guid=UpperCAmelCase , text_a=UpperCAmelCase , text_b=UpperCAmelCase , label=UpperCAmelCase , pairID=UpperCAmelCase ) ) return examples def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> Optional[int]: lowerCamelCase__ : int = {label: i for i, label in enumerate(_UpperCAmelCase )} lowerCamelCase__ : List[Any] = [] for ex_index, example in tqdm.tqdm(enumerate(_UpperCAmelCase ) , desc='convert examples to features' ): if ex_index % 1_0000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCamelCase__ : List[Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=_UpperCAmelCase , max_length=_UpperCAmelCase , padding='max_length' , truncation=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , ) lowerCamelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCamelCase__ : Optional[int] = int(example.pairID ) features.append(InputFeatures(**_UpperCAmelCase , label=_UpperCAmelCase , pairID=_UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features _UpperCAmelCase : str = { """hans""": 3, } _UpperCAmelCase : List[Any] = { """hans""": HansProcessor, }
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCAmelCase : int = {"""configuration_speech_encoder_decoder""": ["""SpeechEncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Tuple = ["""SpeechEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ["""FlaxSpeechEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys _UpperCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[Any] = """▁""" _UpperCAmelCase : Optional[Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = BertGenerationTokenizer UpperCAmelCase__ = False UpperCAmelCase__ = True def A_ ( self : List[Any] ) -> List[str]: super().setUp() lowerCamelCase__ : Dict = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : List[str] = '<s>' lowerCamelCase__ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : List[str] ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(UpperCAmelCase ) , 1002 ) def A_ ( self : List[Any] ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowerCamelCase__ : List[str] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowerCamelCase__ : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def A_ ( self : Dict ) -> Tuple: return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def A_ ( self : Optional[int] ) -> List[str]: lowerCamelCase__ : Union[str, Any] = 'Hello World!' lowerCamelCase__ : Dict = [18536, 2260, 101] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @slow def A_ ( self : Optional[Any] ) -> str: lowerCamelCase__ : List[Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCamelCase__ : Any = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @require_torch @slow def A_ ( self : int ) -> Optional[Any]: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowerCamelCase__ : str = list(self.big_tokenizer.get_vocab().keys() )[:10] lowerCamelCase__ : int = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.big_tokenizer.encode_plus(UpperCAmelCase , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Tuple = BertGenerationConfig() lowerCamelCase__ : Optional[Any] = BertGenerationEncoder(UpperCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase ) model(**UpperCAmelCase ) @slow def A_ ( self : Optional[int] ) -> List[Any]: # fmt: off lowerCamelCase__ : Any = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
50
1
from ..utils import DummyObject, requires_backends class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : List[str] ) -> int: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : int , **UpperCAmelCase : List[str] ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Dict , **UpperCAmelCase : Optional[Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : int , **UpperCAmelCase : Union[str, Any] ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Union[str, Any] ) -> Dict: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Dict ) -> List[str]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : str , **UpperCAmelCase : Optional[Any] ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ) -> Tuple: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Optional[int] ) -> List[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : List[str] , **UpperCAmelCase : List[str] ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : List[str] , **UpperCAmelCase : int ) -> Dict: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Dict ) -> List[str]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Dict , *UpperCAmelCase : List[str] , **UpperCAmelCase : int ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : str ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : str , **UpperCAmelCase : str ) -> List[str]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : int , **UpperCAmelCase : Optional[Any] ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Tuple ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[str] ) -> List[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : int , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[Any] ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : str ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : Dict , **UpperCAmelCase : Union[str, Any] ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : str , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[Any] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : Union[str, Any] ) -> int: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Optional[int] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Dict ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : int , **UpperCAmelCase : List[str] ) -> str: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Any , **UpperCAmelCase : Optional[Any] ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : Tuple , **UpperCAmelCase : Tuple ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : int , **UpperCAmelCase : Dict ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Any ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Tuple ) -> List[str]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Optional[Any] ) -> List[str]: requires_backends(cls , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Any: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> str: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Optional[int]: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Any: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Tuple: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Union[str, Any]: requires_backends(_UpperCAmelCase , ['torch'] ) def SCREAMING_SNAKE_CASE ( *_UpperCAmelCase , **_UpperCAmelCase ) -> Any: requires_backends(_UpperCAmelCase , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> List[str]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Optional[Any] ) -> Dict: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : List[str] ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[str] ) -> Union[str, Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Optional[int] ) -> int: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : List[str] ) -> Tuple: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[int] ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Dict , **UpperCAmelCase : Union[str, Any] ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[str] , *UpperCAmelCase : int , **UpperCAmelCase : Optional[int] ) -> Union[str, Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Any ) -> Tuple: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Any , **UpperCAmelCase : Optional[Any] ) -> int: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : int , *UpperCAmelCase : str , **UpperCAmelCase : List[str] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Dict , **UpperCAmelCase : Union[str, Any] ) -> Tuple: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : int , **UpperCAmelCase : Any ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : Any , **UpperCAmelCase : Any ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Dict , **UpperCAmelCase : List[str] ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Any ) -> Tuple: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Any ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Any , **UpperCAmelCase : str ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Any ) -> Optional[int]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Dict , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Union[str, Any] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : str ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : int ) -> int: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : int , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : List[str] ) -> int: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[int] ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : str , *UpperCAmelCase : List[str] , **UpperCAmelCase : Dict ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : str ) -> Optional[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : Any , **UpperCAmelCase : List[str] ) -> List[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : Optional[int] ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : str , **UpperCAmelCase : Optional[int] ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : str , *UpperCAmelCase : int , **UpperCAmelCase : int ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : int , **UpperCAmelCase : str ) -> Optional[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : List[str] ) -> List[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Any , **UpperCAmelCase : str ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : List[Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : int , *UpperCAmelCase : List[str] , **UpperCAmelCase : Union[str, Any] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : str ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[str] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[str] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Dict , **UpperCAmelCase : int ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : str ) -> str: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[Any] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Tuple , **UpperCAmelCase : Union[str, Any] ) -> int: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Union[str, Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[str] , *UpperCAmelCase : Dict , **UpperCAmelCase : Optional[Any] ) -> str: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Any , **UpperCAmelCase : str ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Optional[Any] ) -> str: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Dict , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Tuple ) -> int: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Any , **UpperCAmelCase : Union[str, Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : List[str] ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Optional[int] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Optional[int] ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Optional[int] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : str , *UpperCAmelCase : int , **UpperCAmelCase : str ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : str , **UpperCAmelCase : Union[str, Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Tuple , **UpperCAmelCase : str ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Any ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : Tuple , **UpperCAmelCase : List[str] ) -> str: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : str ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : int , **UpperCAmelCase : str ) -> Dict: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Optional[int] ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[int] ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : int , **UpperCAmelCase : List[str] ) -> List[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Union[str, Any] ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Dict ) -> int: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : List[Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : str , **UpperCAmelCase : Union[str, Any] ) -> Tuple: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Dict , *UpperCAmelCase : Dict , **UpperCAmelCase : Tuple ) -> Tuple: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : List[str] , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : Optional[Any] ) -> List[str]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : str , **UpperCAmelCase : Optional[Any] ) -> Optional[int]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : int ) -> int: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : str , *UpperCAmelCase : Tuple , **UpperCAmelCase : str ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : int , **UpperCAmelCase : Tuple ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : int , **UpperCAmelCase : List[Any] ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : int ) -> int: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Any , **UpperCAmelCase : Optional[Any] ) -> List[str]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> List[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : str , *UpperCAmelCase : int , **UpperCAmelCase : Any ) -> str: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : Any ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : List[str] ) -> Optional[int]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[Any] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Tuple ) -> Any: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Any ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Optional[int] ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : str , **UpperCAmelCase : int ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Any ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : int , **UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Tuple , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Dict ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : List[Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Dict ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Tuple ) -> Tuple: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : Union[str, Any] ) -> Dict: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Any ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : int , *UpperCAmelCase : str , **UpperCAmelCase : int ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Dict ) -> List[Any]: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Optional[int] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Tuple ) -> Union[str, Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : List[str] ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : List[str] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Union[str, Any] ) -> Tuple: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Dict , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Optional[int] ) -> Union[str, Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : Dict , **UpperCAmelCase : int ) -> int: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : List[str] , *UpperCAmelCase : Dict , **UpperCAmelCase : Dict ) -> List[Any]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[Any] ) -> Optional[Any]: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Dict , *UpperCAmelCase : int , **UpperCAmelCase : Dict ) -> Dict: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Any , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : int ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Optional[int] , *UpperCAmelCase : List[str] , **UpperCAmelCase : Tuple ) -> Any: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Tuple , *UpperCAmelCase : str , **UpperCAmelCase : Dict ) -> Any: requires_backends(cls , ['torch'] ) class lowerCAmelCase ( metaclass=__UpperCamelCase ): UpperCAmelCase__ = ["""torch"""] def __init__( self : Union[str, Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : List[str] ) -> Optional[int]: requires_backends(self , ['torch'] ) @classmethod def A_ ( cls : Union[str, Any] , *UpperCAmelCase : str , **UpperCAmelCase : Optional[int] ) -> str: requires_backends(cls , ['torch'] ) @classmethod def A_ ( cls : Any , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Optional[int] ) -> Optional[Any]: requires_backends(cls , ['torch'] )
50
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _UpperCAmelCase : str = pytest.mark.integration @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : List[Any] ) -> Union[str, Any]: lowerCamelCase__ : List[Any] = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(UpperCAmelCase ) for x in np.arange(30 ).tolist()]} ) return dset def A_ ( self : Optional[Any] ) -> Optional[int]: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() lowerCamelCase__ : List[Any] = dset.map( lambda UpperCAmelCase , UpperCAmelCase : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=UpperCAmelCase , keep_in_memory=UpperCAmelCase ) lowerCamelCase__ : Tuple = dset.add_faiss_index('vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def A_ ( self : Union[str, Any] ) -> int: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : List[str] ) -> Tuple: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : Any ) -> Optional[Any]: lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(UpperCAmelCase , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def A_ ( self : Dict ) -> Dict: from elasticsearch import Elasticsearch lowerCamelCase__ : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : List[Any] = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) lowerCamelCase__ : int = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} lowerCamelCase__ : List[str] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Any ) -> Dict: import faiss lowerCamelCase__ : Tuple = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query lowerCamelCase__ : int = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Any = 1 lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = index.search(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries lowerCamelCase__ : str = np.eye(5 , dtype=np.floataa )[::-1] lowerCamelCase__ , lowerCamelCase__ : Dict = index.search_batch(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search_batch , queries[0] ) lowerCamelCase__ : str = [scores[0] for scores in total_scores] lowerCamelCase__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , UpperCAmelCase ) def A_ ( self : List[Any] ) -> List[Any]: import faiss lowerCamelCase__ : Any = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) lowerCamelCase__ : Tuple = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(UpperCAmelCase ): lowerCamelCase__ : List[str] = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def A_ ( self : List[str] ) -> Optional[int]: import faiss lowerCamelCase__ : Optional[Any] = faiss.IndexFlat(5 ) lowerCamelCase__ : int = FaissIndex(custom_index=UpperCAmelCase ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def A_ ( self : Any ) -> Optional[int]: import faiss lowerCamelCase__ : Any = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: index.save(tmp_file.name ) lowerCamelCase__ : List[Any] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ : List[str] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Tuple = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(UpperCAmelCase ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: import faiss lowerCamelCase__ : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) lowerCamelCase__ : Optional[int] = 'index.faiss' lowerCamelCase__ : Optional[Any] = F"""mock://{index_name}""" index.save(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Tuple = FaissIndex.load(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Optional[int] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Dict = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(_UpperCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Dict ) -> List[Any]: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : Any = Elasticsearch() lowerCamelCase__ : Tuple = {'acknowledged': True} lowerCamelCase__ : Tuple = ElasticSearchIndex(es_client=UpperCAmelCase ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query lowerCamelCase__ : Optional[int] = 'foo' lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search(UpperCAmelCase ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout lowerCamelCase__ : Any = 'foo' lowerCamelCase__ : List[str] = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Tuple = index.search(UpperCAmelCase , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries lowerCamelCase__ : List[str] = ['foo', 'bar', 'foobar'] lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : str = index.search_batch(UpperCAmelCase ) lowerCamelCase__ : List[str] = [scores[0] for scores in total_scores] lowerCamelCase__ : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase ) # batched queries with timeout lowerCamelCase__ : str = ['foo', 'bar', 'foobar'] lowerCamelCase__ : List[Any] = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search_batch(UpperCAmelCase , request_timeout=30 ) lowerCamelCase__ : Optional[Any] = [scores[0] for scores in total_scores] lowerCamelCase__ : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase )
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: if n == 1 or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): return 0 elif n == 2: return 1 else: lowerCamelCase__ : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Dict = 2 while digits < n: index += 1 lowerCamelCase__ : Union[str, Any] = len(str(fibonacci(_UpperCAmelCase ) ) ) return index def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 1000 ) -> int: return fibonacci_digits_index(_UpperCAmelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : str = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCamelCase__ : Tuple = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCamelCase__ : Dict = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCamelCase__ : str = subset[i - 1][j] if arr[i - 1] <= j: lowerCamelCase__ : Dict = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
50
1
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : int = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = XLNetTokenizer UpperCAmelCase__ = XLNetTokenizerFast UpperCAmelCase__ = True UpperCAmelCase__ = True def A_ ( self : List[str] ) -> List[Any]: super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase__ : Optional[Any] = XLNetTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = '<s>' lowerCamelCase__ : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Any: lowerCamelCase__ : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<eod>' ) self.assertEqual(len(UpperCAmelCase ) , 1006 ) def A_ ( self : List[Any] ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A_ ( self : int ) -> Tuple: lowerCamelCase__ : Optional[int] = XLNetTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowerCamelCase__ : Optional[int] = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] ) lowerCamelCase__ : Union[str, Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) lowerCamelCase__ : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = XLNetTokenizer(UpperCAmelCase , do_lower_case=UpperCAmelCase ) lowerCamelCase__ : Tuple = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + '', 'i', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['▁he', 'll', 'o'] ) def A_ ( self : Any ) -> int: lowerCamelCase__ : str = XLNetTokenizer(UpperCAmelCase , do_lower_case=UpperCAmelCase ) lowerCamelCase__ : Any = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) @slow def A_ ( self : List[Any] ) -> Optional[int]: lowerCamelCase__ : List[str] = XLNetTokenizer.from_pretrained('xlnet-base-cased' ) lowerCamelCase__ : List[Any] = tokenizer.encode('sequence builders' , add_special_tokens=UpperCAmelCase ) lowerCamelCase__ : Dict = tokenizer.encode('multi-sequence build' , add_special_tokens=UpperCAmelCase ) lowerCamelCase__ : str = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) lowerCamelCase__ : Any = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def A_ ( self : Optional[Any] ) -> List[str]: # fmt: off lowerCamelCase__ : Tuple = {'input_ids': [[17, 21442, 270, 17, 10, 14645, 318, 34, 17, 4546, 3145, 787, 13, 7752, 22018, 23, 21, 17, 4546, 3145, 787, 13, 3352, 14431, 13, 5500, 11, 1176, 580, 13, 16819, 4797, 23, 17, 10, 17135, 658, 19, 457, 7932, 13, 184, 19, 3154, 17135, 6468, 19, 1404, 12269, 19, 4229, 5356, 16264, 46, 19, 17, 20545, 10395, 9, 9, 9, 11, 28, 6421, 9531, 20729, 17, 10, 353, 17022, 11, 21, 6421, 9531, 16949, 17, 10, 11509, 753, 11, 33, 95, 2421, 7385, 956, 14431, 2626, 25, 842, 7385, 4836, 21, 1429, 2272, 9855, 3120, 161, 24738, 19, 13203, 658, 218, 787, 21, 430, 18482, 847, 2637, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 322, 22178, 27, 1064, 22, 956, 13, 11101, 1429, 5854, 24313, 18953, 40, 422, 24366, 68, 1758, 37, 10483, 14257, 31, 207, 263, 21, 203, 3773, 25, 71, 9735, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 2049, 3442, 17, 13894, 3380, 23, 95, 18, 17634, 2288, 9, 4, 3]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='xlnet-base-cased' , revision='c841166438c31ec7ca9a106dee7bb312b73ae511' , )
50
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """M-CLIP""" def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=1024 , UpperCAmelCase : Tuple=768 , **UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[int] = transformerDimSize lowerCamelCase__ : Optional[Any] = imageDimSize super().__init__(**UpperCAmelCase ) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = MCLIPConfig def __init__( self : List[Any] , UpperCAmelCase : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Dict: super().__init__(UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Tuple = XLMRobertaModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict ) -> Tuple: lowerCamelCase__ : Any = self.transformer(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase )[0] lowerCamelCase__ : int = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(UpperCAmelCase ), embs
50
1
from scipy.stats import pearsonr import datasets _UpperCAmelCase : str = """ Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _UpperCAmelCase : Union[str, Any] = """ Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _UpperCAmelCase : List[Any] = """ @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def A_ ( self : Any ) -> List[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('float' ), 'references': datasets.Value('float' ), } ) , reference_urls=['https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'] , ) def A_ ( self : int , UpperCAmelCase : str , UpperCAmelCase : List[Any] , UpperCAmelCase : Any=False ) -> str: if return_pvalue: lowerCamelCase__ : int = pearsonr(UpperCAmelCase , UpperCAmelCase ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(UpperCAmelCase , UpperCAmelCase )[0] )}
50
from itertools import count def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Optional[Any] = [1] * min_block_length for n in count(_UpperCAmelCase ): fill_count_functions.append(1 ) for block_length in range(_UpperCAmelCase , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from __future__ import annotations from typing import Any class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : float = 0 ) -> None: lowerCamelCase__ , lowerCamelCase__ : str = row, column lowerCamelCase__ : str = [[default_value for c in range(UpperCAmelCase )] for r in range(UpperCAmelCase )] def __str__( self : str ) -> str: lowerCamelCase__ : str = F"""Matrix consist of {self.row} rows and {self.column} columns\n""" # Make string identifier lowerCamelCase__ : Optional[int] = 0 for row_vector in self.array: for obj in row_vector: lowerCamelCase__ : List[Any] = max(UpperCAmelCase , len(str(UpperCAmelCase ) ) ) lowerCamelCase__ : List[Any] = F"""%{max_element_length}s""" # Make string and return def single_line(UpperCAmelCase : list[float] ) -> str: nonlocal string_format_identifier lowerCamelCase__ : List[Any] = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(UpperCAmelCase ) for row_vector in self.array ) return s def __repr__( self : Optional[Any] ) -> str: return str(self ) def A_ ( self : Any , UpperCAmelCase : tuple[int, int] ) -> bool: if not (isinstance(UpperCAmelCase , (list, tuple) ) and len(UpperCAmelCase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Any , UpperCAmelCase : tuple[int, int] ) -> Any: assert self.validate_indicies(UpperCAmelCase ) return self.array[loc[0]][loc[1]] def __setitem__( self : str , UpperCAmelCase : tuple[int, int] , UpperCAmelCase : float ) -> None: assert self.validate_indicies(UpperCAmelCase ) lowerCamelCase__ : Dict = value def __add__( self : Optional[Any] , UpperCAmelCase : Matrix ) -> Matrix: assert isinstance(UpperCAmelCase , UpperCAmelCase ) assert self.row == another.row and self.column == another.column # Add lowerCamelCase__ : List[Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCamelCase__ : Dict = self[r, c] + another[r, c] return result def __neg__( self : Dict ) -> Matrix: lowerCamelCase__ : int = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCamelCase__ : List[str] = -self[r, c] return result def __sub__( self : List[Any] , UpperCAmelCase : Matrix ) -> Matrix: return self + (-another) def __mul__( self : Tuple , UpperCAmelCase : int | float | Matrix ) -> Matrix: if isinstance(UpperCAmelCase , (int, float) ): # Scalar multiplication lowerCamelCase__ : Union[str, Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCamelCase__ : Dict = self[r, c] * another return result elif isinstance(UpperCAmelCase , UpperCAmelCase ): # Matrix multiplication assert self.column == another.row lowerCamelCase__ : Optional[int] = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: lowerCamelCase__ : Any = F"""Unsupported type given for another ({type(UpperCAmelCase )})""" raise TypeError(UpperCAmelCase ) def A_ ( self : Optional[Any] ) -> Matrix: lowerCamelCase__ : str = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): lowerCamelCase__ : Tuple = self[r, c] return result def A_ ( self : Tuple , UpperCAmelCase : Matrix , UpperCAmelCase : Matrix ) -> Any: assert isinstance(UpperCAmelCase , UpperCAmelCase ) and isinstance(UpperCAmelCase , UpperCAmelCase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate lowerCamelCase__ : Any = v.transpose() lowerCamelCase__ : Dict = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( ) -> None: # a^(-1) lowerCamelCase__ : str = Matrix(3 , 3 , 0 ) for i in range(3 ): lowerCamelCase__ : str = 1 print(F"""a^(-1) is {ainv}""" ) # u, v lowerCamelCase__ : Any = Matrix(3 , 1 , 0 ) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple = 1, 2, -3 lowerCamelCase__ : Tuple = Matrix(3 , 1 , 0 ) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = 4, -2, 5 print(F"""u is {u}""" ) print(F"""v is {v}""" ) print(F"""uv^T is {u * v.transpose()}""" ) # Sherman Morrison print(F"""(a + uv^T)^(-1) is {ainv.sherman_morrison(_UpperCAmelCase , _UpperCAmelCase )}""" ) def SCREAMING_SNAKE_CASE ( ) -> None: import doctest doctest.testmod() testa()
50
from __future__ import annotations from typing import Any def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: create_state_space_tree(_UpperCAmelCase , [] , 0 ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> None: if index == len(_UpperCAmelCase ): print(_UpperCAmelCase ) return create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _UpperCAmelCase : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["""A""", """B""", """C"""]) generate_all_subsequences(seq)
50
1
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Tuple = r'\w+[.]\d+' lowerCamelCase__ : List[str] = re.findall(_UpperCAmelCase , _UpperCAmelCase ) for pat in pats: lowerCamelCase__ : str = key.replace(_UpperCAmelCase , '_'.join(pat.split('.' ) ) ) return key def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ : str = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): lowerCamelCase__ : Union[str, Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: lowerCamelCase__ : Any = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: lowerCamelCase__ : str = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer lowerCamelCase__ : Any = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: lowerCamelCase__ : List[str] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer lowerCamelCase__ : Union[str, Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": lowerCamelCase__ : Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight lowerCamelCase__ : int = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias lowerCamelCase__ : Any = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=42 ) -> Optional[int]: # Step 1: Convert pytorch tensor to numpy lowerCamelCase__ : int = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params lowerCamelCase__ : Optional[int] = flax_model.init_weights(PRNGKey(_UpperCAmelCase ) ) lowerCamelCase__ : Union[str, Any] = flatten_dict(_UpperCAmelCase ) lowerCamelCase__ : Optional[int] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowerCamelCase__ : Union[str, Any] = rename_key(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters lowerCamelCase__ , lowerCamelCase__ : List[Any] = rename_key_and_reshape_tensor(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown lowerCamelCase__ : Any = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase )
50
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off _UpperCAmelCase : Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = NllbTokenizer UpperCAmelCase__ = [] UpperCAmelCase__ = [] def __init__( self : Tuple , UpperCAmelCase : int=None , UpperCAmelCase : Any=None , UpperCAmelCase : str="<s>" , UpperCAmelCase : Optional[Any]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Tuple="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[str]="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=False , **UpperCAmelCase : Optional[int] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token lowerCamelCase__ : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase , tokenizer_file=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , src_lang=UpperCAmelCase , tgt_lang=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , legacy_behaviour=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = vocab_file lowerCamelCase__ : Dict = False if not self.vocab_file else True lowerCamelCase__ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCamelCase__ : str = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : int = src_lang if src_lang is not None else 'eng_Latn' lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A_ ( self : int ) -> str: return self._src_lang @src_lang.setter def A_ ( self : List[Any] , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] , UpperCAmelCase : Optional[str] , **UpperCAmelCase : List[str] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCamelCase__ : Optional[int] = src_lang lowerCamelCase__ : Optional[int] = self(UpperCAmelCase , add_special_tokens=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.convert_tokens_to_ids(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tgt_lang_id return inputs def A_ ( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : str = "eng_Latn" , UpperCAmelCase : Optional[List[str]] = None , UpperCAmelCase : str = "fra_Latn" , **UpperCAmelCase : Dict , ) -> BatchEncoding: lowerCamelCase__ : Any = src_lang lowerCamelCase__ : int = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Optional[int]: return self.set_src_lang_special_tokens(self.src_lang ) def A_ ( self : Any ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A_ ( self : str , UpperCAmelCase : Optional[Any] ) -> None: lowerCamelCase__ : int = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : int = [self.cur_lang_code] lowerCamelCase__ : Tuple = [self.eos_token_id] lowerCamelCase__ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : int , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Union[str, Any] = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : Dict = [] lowerCamelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : Any = [self.cur_lang_code] lowerCamelCase__ : Optional[Any] = [self.eos_token_id] lowerCamelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCamelCase__ : int = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
1
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer _UpperCAmelCase : List[str] = logging.get_logger(__name__) _UpperCAmelCase : Tuple = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : str = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } _UpperCAmelCase : Union[str, Any] = { """allenai/led-base-16384""": 1_63_84, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = LEDTokenizer UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : int , UpperCAmelCase : Any=None , UpperCAmelCase : Tuple=None , UpperCAmelCase : Dict=None , UpperCAmelCase : int="replace" , UpperCAmelCase : int="<s>" , UpperCAmelCase : Optional[int]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : List[Any]="<s>" , UpperCAmelCase : Any="<unk>" , UpperCAmelCase : str="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : str=False , UpperCAmelCase : Dict=True , **UpperCAmelCase : Optional[int] , ) -> List[Any]: super().__init__( UpperCAmelCase , UpperCAmelCase , tokenizer_file=UpperCAmelCase , errors=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , add_prefix_space=UpperCAmelCase , trim_offsets=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : Union[str, Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , UpperCAmelCase ) != add_prefix_space: lowerCamelCase__ : int = getattr(UpperCAmelCase , pre_tok_state.pop('type' ) ) lowerCamelCase__ : Optional[int] = add_prefix_space lowerCamelCase__ : List[str] = pre_tok_class(**UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCamelCase__ : str = 'post_processor' lowerCamelCase__ : Tuple = getattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) if tokenizer_component_instance: lowerCamelCase__ : Any = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : int = tuple(state['sep'] ) if "cls" in state: lowerCamelCase__ : Optional[int] = tuple(state['cls'] ) lowerCamelCase__ : Optional[int] = False if state.get('add_prefix_space' , UpperCAmelCase ) != add_prefix_space: lowerCamelCase__ : Dict = add_prefix_space lowerCamelCase__ : Optional[Any] = True if state.get('trim_offsets' , UpperCAmelCase ) != trim_offsets: lowerCamelCase__ : Any = trim_offsets lowerCamelCase__ : Tuple = True if changes_to_apply: lowerCamelCase__ : Tuple = getattr(UpperCAmelCase , state.pop('type' ) ) lowerCamelCase__ : Any = component_class(**UpperCAmelCase ) setattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A_ ( self : str ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def A_ ( self : Dict , UpperCAmelCase : Tuple ) -> str: lowerCamelCase__ : Any = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else value lowerCamelCase__ : Union[str, Any] = value def A_ ( self : Any , *UpperCAmelCase : Tuple , **UpperCAmelCase : List[str] ) -> BatchEncoding: lowerCamelCase__ : Union[str, Any] = kwargs.get('is_split_into_words' , UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : List[Any] , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ) -> BatchEncoding: lowerCamelCase__ : Optional[Any] = kwargs.get('is_split_into_words' , UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : Union[str, Any] = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase ) def A_ ( self : Dict , UpperCAmelCase : Dict , UpperCAmelCase : str=None ) -> str: lowerCamelCase__ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A_ ( self : Dict , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : Dict , UpperCAmelCase : Union[Dict[str, EncodedInput], BatchEncoding] , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Optional[bool] = None , ) -> dict: lowerCamelCase__ : List[str] = super()._pad( encoded_inputs=UpperCAmelCase , max_length=UpperCAmelCase , padding_strategy=UpperCAmelCase , pad_to_multiple_of=UpperCAmelCase , return_attention_mask=UpperCAmelCase , ) # Load from model defaults if return_attention_mask is None: lowerCamelCase__ : Tuple = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCamelCase__ : List[Any] = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCamelCase__ : Dict = len(encoded_inputs['global_attention_mask'] ) != len(UpperCAmelCase ) if needs_to_be_padded: lowerCamelCase__ : Any = len(UpperCAmelCase ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCamelCase__ : List[str] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowerCamelCase__ : Dict = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
50
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ , lowerCamelCase__ : List[str] = emb.weight.shape lowerCamelCase__ : Tuple = nn.Linear(_UpperCAmelCase , _UpperCAmelCase , bias=_UpperCAmelCase ) lowerCamelCase__ : Dict = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location='cpu' ) lowerCamelCase__ : List[str] = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCamelCase__ : Optional[int] = mam_aaa['model'] remove_ignore_keys_(_UpperCAmelCase ) lowerCamelCase__ : str = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCamelCase__ : Union[str, Any] = MaMaaaConfig( vocab_size=_UpperCAmelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCamelCase__ : Optional[Any] = state_dict['decoder.embed_tokens.weight'] lowerCamelCase__ : Union[str, Any] = MaMaaaForConditionalGeneration(_UpperCAmelCase ) model.model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) lowerCamelCase__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase : str = parser.parse_args() _UpperCAmelCase : Optional[Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
50
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = CpmAntTokenizer UpperCAmelCase__ = False def A_ ( self : List[str] ) -> Union[str, Any]: super().setUp() lowerCamelCase__ : List[Any] = [ '<d>', '</d>', '<s>', '</s>', '</_>', '<unk>', '<pad>', '</n>', '我', '是', 'C', 'P', 'M', 'A', 'n', 't', ] lowerCamelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) @tooslow def A_ ( self : Union[str, Any] ) -> Tuple: lowerCamelCase__ : int = CpmAntTokenizer.from_pretrained('openbmb/cpm-ant-10b' ) lowerCamelCase__ : Any = '今天天气真好!' lowerCamelCase__ : Dict = ['今天', '天气', '真', '好', '!'] lowerCamelCase__ : Any = tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : List[str] = '今天天气真好!' lowerCamelCase__ : List[str] = [tokenizer.bos_token] + tokens lowerCamelCase__ : int = [6, 9802, 14962, 2082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , UpperCAmelCase ) lowerCamelCase__ : str = tokenizer.decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
1
import os def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = "matrix.txt" ) -> int: with open(os.path.join(os.path.dirname(_UpperCAmelCase ) , _UpperCAmelCase ) ) as in_file: lowerCamelCase__ : int = in_file.read() lowerCamelCase__ : Tuple = [[int(_UpperCAmelCase ) for cell in row.split(',' )] for row in data.strip().splitlines()] lowerCamelCase__ : Any = [[0 for cell in row] for row in grid] lowerCamelCase__ : Optional[int] = len(grid[0] ) lowerCamelCase__ : Optional[Any] = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] lowerCamelCase__ : Any = grid[0][0] for i in range(1 , _UpperCAmelCase ): lowerCamelCase__ : Any = grid[0][i] + dp[0][i - 1] for i in range(1 , _UpperCAmelCase ): lowerCamelCase__ : Optional[int] = grid[i][0] + dp[i - 1][0] for i in range(1 , _UpperCAmelCase ): for j in range(1 , _UpperCAmelCase ): lowerCamelCase__ : List[Any] = grid[i][j] + min(dp[i - 1][j] , dp[i][j - 1] ) return dp[-1][-1] if __name__ == "__main__": print(F"""{solution() = }""")
50
import flax.linen as nn import jax import jax.numpy as jnp class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , UpperCAmelCase : Dict ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = hidden_states.shape lowerCamelCase__ : Union[str, Any] = jax.image.resize( UpperCAmelCase , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : str , UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = None UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels lowerCamelCase__ : Tuple = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : int = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Union[str, Any] = nn.Dense(UpperCAmelCase , dtype=self.dtype ) lowerCamelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : List[Any] = nn.Dropout(self.dropout_prob ) lowerCamelCase__ : Tuple = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Optional[Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCamelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCamelCase__ : Dict = nn.Conv( UpperCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=True ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = hidden_states lowerCamelCase__ : List[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[Any] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Any = self.conva(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.time_emb_proj(nn.swish(UpperCAmelCase ) ) lowerCamelCase__ : List[str] = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase , 1 ) , 1 ) lowerCamelCase__ : List[str] = hidden_states + temb lowerCamelCase__ : Optional[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[str] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.dropout(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : str = self.conva(UpperCAmelCase ) if self.conv_shortcut is not None: lowerCamelCase__ : Dict = self.conv_shortcut(UpperCAmelCase ) return hidden_states + residual
50
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } _UpperCAmelCase : Any = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]="[UNK]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Tuple="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=None , **UpperCAmelCase : Optional[int] , ) -> str: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : str = strip_accents lowerCamelCase__ : Optional[Any] = tokenize_chinese_chars lowerCamelCase__ : int = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : str = do_lower_case def A_ ( self : Optional[int] , UpperCAmelCase : int , **UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH lowerCamelCase__ : Optional[int] = text lowerCamelCase__ : Dict = kwargs.pop('text_pair' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = kwargs.pop('return_tensors' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: lowerCamelCase__ : Tuple = batch_text_pair[idx] else: lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Any = encoded_candidates.get('input_ids' ) lowerCamelCase__ : Union[str, Any] = encoded_candidates.get('attention_mask' ) lowerCamelCase__ : Tuple = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) lowerCamelCase__ : int = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ) -> List[str]: lowerCamelCase__ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Optional[Any] = set() # edges = list of graph's edges lowerCamelCase__ : List[str] = get_edges(_UpperCAmelCase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCamelCase__ , lowerCamelCase__ : str = edges.pop() chosen_vertices.add(_UpperCAmelCase ) chosen_vertices.add(_UpperCAmelCase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_UpperCAmelCase ) return chosen_vertices def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Union[str, Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
50
1
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline _UpperCAmelCase : Tuple = datasets.utils.logging.get_logger(__name__) @dataclass class lowerCAmelCase ( datasets.BuilderConfig ): UpperCAmelCase__ = None UpperCAmelCase__ = "utf-8" UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = True # deprecated UpperCAmelCase__ = None # deprecated UpperCAmelCase__ = 10 << 20 # 10MB UpperCAmelCase__ = None class lowerCAmelCase ( datasets.ArrowBasedBuilder ): UpperCAmelCase__ = JsonConfig def A_ ( self : Tuple ) -> Optional[int]: if self.config.block_size is not None: logger.warning('The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead' ) lowerCamelCase__ : Optional[Any] = self.config.block_size if self.config.use_threads is not True: logger.warning( 'The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.' ) if self.config.newlines_in_values is not None: raise ValueError('The JSON loader parameter `newlines_in_values` is no longer supported' ) return datasets.DatasetInfo(features=self.config.features ) def A_ ( self : Any , UpperCAmelCase : List[str] ) -> Optional[int]: if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) lowerCamelCase__ : str = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCAmelCase , (str, list, tuple) ): lowerCamelCase__ : Optional[Any] = data_files if isinstance(UpperCAmelCase , UpperCAmelCase ): lowerCamelCase__ : List[Any] = [files] lowerCamelCase__ : Optional[Any] = [dl_manager.iter_files(UpperCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files} )] lowerCamelCase__ : Any = [] for split_name, files in data_files.items(): if isinstance(UpperCAmelCase , UpperCAmelCase ): lowerCamelCase__ : Tuple = [files] lowerCamelCase__ : Any = [dl_manager.iter_files(UpperCAmelCase ) for file in files] splits.append(datasets.SplitGenerator(name=UpperCAmelCase , gen_kwargs={'files': files} ) ) return splits def A_ ( self : Any , UpperCAmelCase : pa.Table ) -> pa.Table: if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): lowerCamelCase__ : str = self.config.features.arrow_schema.field(UpperCAmelCase ).type lowerCamelCase__ : int = pa_table.append_column(UpperCAmelCase , pa.array([None] * len(UpperCAmelCase ) , type=UpperCAmelCase ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example lowerCamelCase__ : int = table_cast(UpperCAmelCase , self.config.features.arrow_schema ) return pa_table def A_ ( self : str , UpperCAmelCase : Tuple ) -> Dict: for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCAmelCase ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(UpperCAmelCase , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCamelCase__ : Optional[Any] = json.load(UpperCAmelCase ) # We keep only the field we are interested in lowerCamelCase__ : Optional[int] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(UpperCAmelCase , (list, tuple) ): lowerCamelCase__ : Any = set().union(*[row.keys() for row in dataset] ) lowerCamelCase__ : List[Any] = {col: [row.get(UpperCAmelCase ) for row in dataset] for col in keys} else: lowerCamelCase__ : int = dataset lowerCamelCase__ : Union[str, Any] = pa.Table.from_pydict(UpperCAmelCase ) yield file_idx, self._cast_table(UpperCAmelCase ) # If the file has one json object per line else: with open(UpperCAmelCase , 'rb' ) as f: lowerCamelCase__ : Any = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small lowerCamelCase__ : Optional[int] = max(self.config.chunksize // 32 , 16 << 10 ) lowerCamelCase__ : List[str] = ( self.config.encoding_errors if self.config.encoding_errors is not None else 'strict' ) while True: lowerCamelCase__ : Optional[int] = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(UpperCAmelCase ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": lowerCamelCase__ : Union[str, Any] = batch.decode(self.config.encoding , errors=UpperCAmelCase ).encode('utf-8' ) try: while True: try: lowerCamelCase__ : List[str] = paj.read_json( io.BytesIO(UpperCAmelCase ) , read_options=paj.ReadOptions(block_size=UpperCAmelCase ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(UpperCAmelCase , pa.ArrowInvalid ) and "straddling" not in str(UpperCAmelCase ) or block_size > len(UpperCAmelCase ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F"""Batch of {len(UpperCAmelCase )} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.""" ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( UpperCAmelCase , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: lowerCamelCase__ : str = json.load(UpperCAmelCase ) except json.JSONDecodeError: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase )}: {e}""" ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(UpperCAmelCase , UpperCAmelCase ): # list is the only sequence type supported in JSON try: lowerCamelCase__ : List[str] = set().union(*[row.keys() for row in dataset] ) lowerCamelCase__ : Optional[Any] = {col: [row.get(UpperCAmelCase ) for row in dataset] for col in keys} lowerCamelCase__ : int = pa.Table.from_pydict(UpperCAmelCase ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase )}: {e}""" ) raise ValueError(F"""Not able to read records in the JSON file at {file}.""" ) from None yield file_idx, self._cast_table(UpperCAmelCase ) break else: logger.error(F"""Failed to read file '{file}' with error {type(UpperCAmelCase )}: {e}""" ) raise ValueError( F"""Not able to read records in the JSON file at {file}. """ F"""You should probably indicate the field of the JSON file containing your records. """ F"""This JSON file contain the following fields: {str(list(dataset.keys() ) )}. """ F"""Select the correct one and provide it as `field='XXX'` to the dataset loading method. """ ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(UpperCAmelCase ) batch_idx += 1
50
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
1
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _UpperCAmelCase : int = logging.getLogger() def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : List[str] = argparse.ArgumentParser() parser.add_argument('-f' ) lowerCamelCase__ : Tuple = parser.parse_args() return args.f class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Optional[int] ) -> None: lowerCamelCase__ : List[str] = logging.StreamHandler(sys.stdout ) logger.addHandler(UpperCAmelCase ) def A_ ( self : List[str] , UpperCAmelCase : Dict ) -> List[str]: lowerCamelCase__ : Any = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , 'run_glue_deebert.py' ) with patch.object(UpperCAmelCase , 'argv' , UpperCAmelCase ): lowerCamelCase__ : Optional[int] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(UpperCAmelCase , 0.6_6_6 ) @slow @require_torch_non_multi_gpu def A_ ( self : str ) -> Optional[int]: lowerCamelCase__ : Any = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(UpperCAmelCase ) lowerCamelCase__ : List[Any] = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(UpperCAmelCase )
50
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
1
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : List[str] = { """vocab_file""": """vocab.txt""", """merges_file""": """bpe.codes""", } _UpperCAmelCase : Tuple = { """vocab_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt""", }, """merges_file""": { """vinai/phobert-base""": """https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes""", """vinai/phobert-large""": """https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes""", }, } _UpperCAmelCase : List[str] = { """vinai/phobert-base""": 2_56, """vinai/phobert-large""": 2_56, } def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = set() lowerCamelCase__ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCamelCase__ : str = char lowerCamelCase__ : List[Any] = set(_UpperCAmelCase ) return pairs class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Any , UpperCAmelCase : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : List[str]="<s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Dict="</s>" , UpperCAmelCase : List[str]="<s>" , UpperCAmelCase : Optional[int]="<unk>" , UpperCAmelCase : Any="<pad>" , UpperCAmelCase : int="<mask>" , **UpperCAmelCase : Tuple , ) -> List[Any]: super().__init__( bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : Union[str, Any] = vocab_file lowerCamelCase__ : int = merges_file lowerCamelCase__ : List[Any] = {} lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : str = 1 lowerCamelCase__ : Optional[int] = 2 lowerCamelCase__ : str = 3 self.add_from_file(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(UpperCAmelCase , encoding='utf-8' ) as merges_handle: lowerCamelCase__ : Optional[Any] = merges_handle.read().split('\n' )[:-1] lowerCamelCase__ : List[str] = [tuple(merge.split()[:-1] ) for merge in merges] lowerCamelCase__ : str = dict(zip(UpperCAmelCase , range(len(UpperCAmelCase ) ) ) ) lowerCamelCase__ : str = {} def A_ ( self : List[str] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCamelCase__ : Any = [self.cls_token_id] lowerCamelCase__ : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None , UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase , token_ids_a=UpperCAmelCase , already_has_special_tokens=UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase )) + [1] return [1] + ([0] * len(UpperCAmelCase )) + [1, 1] + ([0] * len(UpperCAmelCase )) + [1] def A_ ( self : Union[str, Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Optional[int] = [self.sep_token_id] lowerCamelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A_ ( self : Dict ) -> Any: return len(self.encoder ) def A_ ( self : int ) -> Tuple: return dict(self.encoder , **self.added_tokens_encoder ) def A_ ( self : Any , UpperCAmelCase : Optional[Any] ) -> Tuple: if token in self.cache: return self.cache[token] lowerCamelCase__ : Union[str, Any] = tuple(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) lowerCamelCase__ : Dict = get_pairs(UpperCAmelCase ) if not pairs: return token while True: lowerCamelCase__ : Optional[int] = min(UpperCAmelCase , key=lambda UpperCAmelCase : self.bpe_ranks.get(UpperCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowerCamelCase__ , lowerCamelCase__ : Dict = bigram lowerCamelCase__ : Dict = [] lowerCamelCase__ : str = 0 while i < len(UpperCAmelCase ): try: lowerCamelCase__ : Dict = word.index(UpperCAmelCase , UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCamelCase__ : Any = j if word[i] == first and i < len(UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCamelCase__ : str = tuple(UpperCAmelCase ) lowerCamelCase__ : List[Any] = new_word if len(UpperCAmelCase ) == 1: break else: lowerCamelCase__ : int = get_pairs(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = '@@ '.join(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = word[:-4] lowerCamelCase__ : Any = word return word def A_ ( self : int , UpperCAmelCase : List[str] ) -> int: lowerCamelCase__ : str = [] lowerCamelCase__ : List[str] = re.findall(R'\S+\n?' , UpperCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(UpperCAmelCase ).split(' ' ) ) ) return split_tokens def A_ ( self : str , UpperCAmelCase : Any ) -> Optional[Any]: return self.encoder.get(UpperCAmelCase , self.encoder.get(self.unk_token ) ) def A_ ( self : Dict , UpperCAmelCase : Dict ) -> Optional[int]: return self.decoder.get(UpperCAmelCase , self.unk_token ) def A_ ( self : str , UpperCAmelCase : Union[str, Any] ) -> Dict: lowerCamelCase__ : Optional[int] = ' '.join(UpperCAmelCase ).replace('@@ ' , '' ).strip() return out_string def A_ ( self : Any , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : Optional[int] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowerCamelCase__ : Optional[int] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) if os.path.abspath(self.merges_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.merges_file , UpperCAmelCase ) return out_vocab_file, out_merge_file def A_ ( self : List[str] , UpperCAmelCase : str ) -> Optional[int]: if isinstance(UpperCAmelCase , UpperCAmelCase ): try: with open(UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F"""Incorrect encoding detected in {f}, please rebuild the dataset""" ) return lowerCamelCase__ : Tuple = f.readlines() for lineTmp in lines: lowerCamelCase__ : Dict = lineTmp.strip() lowerCamelCase__ : str = line.rfind(' ' ) if idx == -1: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt>\'' ) lowerCamelCase__ : str = line[:idx] lowerCamelCase__ : int = len(self.encoder )
50
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : str=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : str=9 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=True , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : List[str]=5 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Union[str, Any]=37 , UpperCAmelCase : int=8 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=0.0_0_2 , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=0 , UpperCAmelCase : Union[str, Any]=0 , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , ) -> Union[str, Any]: lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = encoder_seq_length lowerCamelCase__ : int = decoder_seq_length # For common tests lowerCamelCase__ : List[str] = self.decoder_seq_length lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : List[Any] = use_attention_mask lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = d_ff lowerCamelCase__ : Optional[Any] = relative_attention_num_buckets lowerCamelCase__ : Any = dropout_rate lowerCamelCase__ : Any = initializer_factor lowerCamelCase__ : Union[str, Any] = eos_token_id lowerCamelCase__ : List[str] = pad_token_id lowerCamelCase__ : List[str] = decoder_start_token_id lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = decoder_layers def A_ ( self : List[Any] ) -> int: return TaConfig.from_pretrained('google/umt5-base' ) def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=None , ) -> List[str]: if attention_mask is None: lowerCamelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCamelCase__ : Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCamelCase__ : int = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase ) if decoder_head_mask is None: lowerCamelCase__ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A_ ( self : str ) -> List[str]: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCamelCase__ : List[str] = input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Union[str, Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Dict = self.get_config() lowerCamelCase__ : Tuple = config.num_attention_heads lowerCamelCase__ : Any = self.prepare_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, input_dict def A_ ( self : Tuple ) -> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Dict = self.prepare_config_and_inputs() return config, inputs_dict def A_ ( self : Optional[int] ) -> List[str]: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Union[str, Any] ) -> Dict: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Dict , ) -> str: lowerCamelCase__ : Dict = UMTaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , ) lowerCamelCase__ : Any = model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) lowerCamelCase__ : Dict = result.last_hidden_state lowerCamelCase__ : Any = result.past_key_values lowerCamelCase__ : List[Any] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[Any] , ) -> Optional[int]: lowerCamelCase__ : List[Any] = UMTaModel(config=UpperCAmelCase ).get_decoder().to(UpperCAmelCase ).eval() # first forward pass lowerCamelCase__ : Tuple = model(UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , use_cache=UpperCAmelCase ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) + 1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCamelCase__ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )['last_hidden_state'] lowerCamelCase__ : str = model(UpperCAmelCase , past_key_values=UpperCAmelCase )['last_hidden_state'] # select random slice lowerCamelCase__ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase__ : Tuple = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase__ : List[str] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase , UpperCAmelCase , atol=1e-3 ) ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = UMTaModel(config=UpperCAmelCase ).to(UpperCAmelCase ).half().eval() lowerCamelCase__ : Optional[int] = model(**UpperCAmelCase )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCAmelCase ).any().item() ) @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCAmelCase__ = (UMTaForConditionalGeneration,) if is_torch_available() else () UpperCAmelCase__ = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = True UpperCAmelCase__ = True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCAmelCase__ = [0.8, 0.9] def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Tuple = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCAmelCase , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: lowerCamelCase__ : int = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Any = config_and_inputs[0] lowerCamelCase__ : Any = UMTaForConditionalGeneration(UpperCAmelCase ).eval() model.to(UpperCAmelCase ) lowerCamelCase__ : Tuple = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), } for attn_name, (name, mask) in zip(UpperCAmelCase , head_masking.items() ): lowerCamelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCamelCase__ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ) lowerCamelCase__ : Tuple = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase , return_dict_in_generate=UpperCAmelCase , **UpperCAmelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCamelCase__ : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=UpperCAmelCase ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=UpperCAmelCase , legacy=UpperCAmelCase ) lowerCamelCase__ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowerCamelCase__ : Tuple = tokenizer(UpperCAmelCase , return_tensors='pt' , padding=UpperCAmelCase ).input_ids # fmt: off lowerCamelCase__ : Any = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model.generate(input_ids.to(UpperCAmelCase ) ) lowerCamelCase__ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowerCamelCase__ : Union[str, Any] = tokenizer.batch_decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
1
_UpperCAmelCase : Dict = [ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
class lowerCAmelCase : def __init__( self : str ) -> Optional[Any]: lowerCamelCase__ : Tuple = {} def A_ ( self : List[Any] ) -> None: print(self.vertex ) for i in self.vertex: print(UpperCAmelCase , ' -> ' , ' -> '.join([str(UpperCAmelCase ) for j in self.vertex[i]] ) ) def A_ ( self : List[Any] , UpperCAmelCase : int , UpperCAmelCase : int ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(UpperCAmelCase ) else: # else make a new vertex lowerCamelCase__ : Union[str, Any] = [to_vertex] def A_ ( self : Optional[int] ) -> None: # visited array for storing already visited nodes lowerCamelCase__ : List[Any] = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(UpperCAmelCase , UpperCAmelCase ) def A_ ( self : Dict , UpperCAmelCase : int , UpperCAmelCase : list ) -> None: # mark start vertex as visited lowerCamelCase__ : int = True print(UpperCAmelCase , end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(UpperCAmelCase , UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : int = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
50
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase ): @register_to_config def __init__( self : List[str] , UpperCAmelCase : int = 65536 , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 0 , UpperCAmelCase : str = "fourier" , UpperCAmelCase : bool = True , UpperCAmelCase : bool = False , UpperCAmelCase : float = 0.0 , UpperCAmelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase : str = None , UpperCAmelCase : Tuple[int] = (32, 32, 64) , UpperCAmelCase : str = None , UpperCAmelCase : int = 8 , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = False , ) -> List[Any]: super().__init__() lowerCamelCase__ : Optional[int] = sample_size # time if time_embedding_type == "fourier": lowerCamelCase__ : Optional[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase , log=UpperCAmelCase , flip_sin_to_cos=UpperCAmelCase ) lowerCamelCase__ : Any = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCamelCase__ : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase , downscale_freq_shift=UpperCAmelCase ) lowerCamelCase__ : Dict = block_out_channels[0] if use_timestep_embedding: lowerCamelCase__ : str = block_out_channels[0] * 4 lowerCamelCase__ : List[Any] = TimestepEmbedding( in_channels=UpperCAmelCase , time_embed_dim=UpperCAmelCase , act_fn=UpperCAmelCase , out_dim=block_out_channels[0] , ) lowerCamelCase__ : Any = nn.ModuleList([] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Optional[int] = None # down lowerCamelCase__ : Optional[int] = in_channels for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = output_channel lowerCamelCase__ : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Optional[int] = get_down_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Optional[int] = get_mid_block( UpperCAmelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase , add_downsample=UpperCAmelCase , ) # up lowerCamelCase__ : Optional[int] = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = reversed_block_out_channels[0] if out_block_type is None: lowerCamelCase__ : List[str] = out_channels else: lowerCamelCase__ : Any = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Tuple = output_channel lowerCamelCase__ : Union[str, Any] = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase ) - 1 else final_upsample_channels ) lowerCamelCase__ : List[str] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Dict = get_up_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : int = output_channel # out lowerCamelCase__ : int = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCamelCase__ : List[Any] = get_out_block( out_block_type=UpperCAmelCase , num_groups_out=UpperCAmelCase , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase , act_fn=UpperCAmelCase , fc_dim=block_out_channels[-1] // 4 , ) def A_ ( self : List[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Union[torch.Tensor, float, int] , UpperCAmelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: lowerCamelCase__ : Optional[Any] = timestep if not torch.is_tensor(UpperCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(sample.device ) lowerCamelCase__ : Optional[int] = self.time_proj(UpperCAmelCase ) if self.config.use_timestep_embedding: lowerCamelCase__ : str = self.time_mlp(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = timestep_embed[..., None] lowerCamelCase__ : str = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCamelCase__ : str = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCamelCase__ : str = () for downsample_block in self.down_blocks: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = downsample_block(hidden_states=UpperCAmelCase , temb=UpperCAmelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCamelCase__ : Optional[Any] = self.mid_block(UpperCAmelCase , UpperCAmelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCamelCase__ : Dict = down_block_res_samples[-1:] lowerCamelCase__ : Optional[Any] = down_block_res_samples[:-1] lowerCamelCase__ : Any = upsample_block(UpperCAmelCase , res_hidden_states_tuple=UpperCAmelCase , temb=UpperCAmelCase ) # 5. post-process if self.out_block: lowerCamelCase__ : Any = self.out_block(UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase )
50
1
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : int = R""" Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. """ class lowerCAmelCase ( __UpperCamelCase ): @add_start_docstrings(UpperCAmelCase ) def __call__( self : Optional[Any] , UpperCAmelCase : torch.LongTensor , UpperCAmelCase : torch.FloatTensor , **UpperCAmelCase : List[Any] ) -> bool: raise NotImplementedError('StoppingCriteria needs to be subclassed' ) class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : List[Any] , UpperCAmelCase : int , UpperCAmelCase : Optional[int] = None ) -> Optional[Any]: lowerCamelCase__ : List[str] = max_length lowerCamelCase__ : Optional[Any] = max_position_embeddings @add_start_docstrings(UpperCAmelCase ) def __call__( self : Optional[Any] , UpperCAmelCase : torch.LongTensor , UpperCAmelCase : torch.FloatTensor , **UpperCAmelCase : Optional[Any] ) -> bool: lowerCamelCase__ : Optional[int] = input_ids.shape[-1] lowerCamelCase__ : int = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( 'This is a friendly reminder - the current text generation call will exceed the model\'s predefined ' F"""maximum length ({self.max_position_embeddings}). Depending on the model, you may observe """ 'exceptions, performance degradation, or nothing at all.' ) return is_done class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[Any] , UpperCAmelCase : int , UpperCAmelCase : int ) -> Dict: warnings.warn( 'The class `MaxNewTokensCriteria` is deprecated. ' F"""Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` """ 'with `max_length = start_length + max_new_tokens` instead.' , UpperCAmelCase , ) lowerCamelCase__ : Dict = start_length lowerCamelCase__ : str = max_new_tokens lowerCamelCase__ : Optional[Any] = start_length + max_new_tokens @add_start_docstrings(UpperCAmelCase ) def __call__( self : List[Any] , UpperCAmelCase : torch.LongTensor , UpperCAmelCase : torch.FloatTensor , **UpperCAmelCase : Optional[Any] ) -> bool: return input_ids.shape[-1] >= self.max_length class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Union[str, Any] , UpperCAmelCase : float , UpperCAmelCase : Optional[float] = None ) -> List[str]: lowerCamelCase__ : Union[str, Any] = max_time lowerCamelCase__ : int = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(UpperCAmelCase ) def __call__( self : int , UpperCAmelCase : torch.LongTensor , UpperCAmelCase : torch.FloatTensor , **UpperCAmelCase : str ) -> bool: return time.time() - self.initial_timestamp > self.max_time class lowerCAmelCase ( __UpperCamelCase ): @add_start_docstrings(UpperCAmelCase ) def __call__( self : Dict , UpperCAmelCase : torch.LongTensor , UpperCAmelCase : torch.FloatTensor , **UpperCAmelCase : int ) -> bool: return any(criteria(UpperCAmelCase , UpperCAmelCase ) for criteria in self ) @property def A_ ( self : str ) -> Optional[int]: for stopping_criterium in self: if isinstance(UpperCAmelCase , UpperCAmelCase ): return stopping_criterium.max_length elif isinstance(UpperCAmelCase , UpperCAmelCase ): return stopping_criterium.max_length return None def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> StoppingCriteriaList: lowerCamelCase__ : Union[str, Any] = stopping_criteria.max_length lowerCamelCase__ : List[str] = deepcopy(_UpperCAmelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('You set different `max_length` for stopping criteria and `max_length` parameter' , _UpperCAmelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=_UpperCAmelCase ) ) return new_stopping_criteria
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
import numpy # List of input, output pairs _UpperCAmelCase : Union[str, Any] = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) _UpperCAmelCase : Dict = (((5_15, 22, 13), 5_55), ((61, 35, 49), 1_50)) _UpperCAmelCase : Tuple = [2, 4, 1, 5] _UpperCAmelCase : Optional[Any] = len(train_data) _UpperCAmelCase : Optional[Any] = 0.009 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase="train" ) -> List[str]: return calculate_hypothesis_value(_UpperCAmelCase , _UpperCAmelCase ) - output( _UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : str = 0 for i in range(len(_UpperCAmelCase ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=m ) -> Dict: lowerCamelCase__ : Union[str, Any] = 0 for i in range(_UpperCAmelCase ): if index == -1: summation_value += _error(_UpperCAmelCase ) else: summation_value += _error(_UpperCAmelCase ) * train_data[i][0][index] return summation_value def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : List[Any] = summation_of_cost_derivative(_UpperCAmelCase , _UpperCAmelCase ) / m return cost_derivative_value def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: global parameter_vector # Tune these values to set a tolerance value for predicted output lowerCamelCase__ : List[str] = 0.000_002 lowerCamelCase__ : Any = 0 lowerCamelCase__ : Tuple = 0 while True: j += 1 lowerCamelCase__ : str = [0, 0, 0, 0] for i in range(0 , len(_UpperCAmelCase ) ): lowerCamelCase__ : Optional[Any] = get_cost_derivative(i - 1 ) lowerCamelCase__ : Tuple = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( _UpperCAmelCase , _UpperCAmelCase , atol=_UpperCAmelCase , rtol=_UpperCAmelCase , ): break lowerCamelCase__ : str = temp_parameter_vector print(('Number of iterations:', j) ) def SCREAMING_SNAKE_CASE ( ) -> Optional[int]: for i in range(len(_UpperCAmelCase ) ): print(('Actual output value:', output(_UpperCAmelCase , 'test' )) ) print(('Hypothesis output:', calculate_hypothesis_value(_UpperCAmelCase , 'test' )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
50
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[int] = [] embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", F"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", F"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", F"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", F"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Tuple = [] attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", F"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", F"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", F"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", F"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] token.append((F"""cvt.encoder.stages.{idx}.cls_token""", 'stage2.cls_token') ) return token def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Tuple = 'imagenet-1k-id2label.json' lowerCamelCase__ : Union[str, Any] = 1000 lowerCamelCase__ : Optional[Any] = 'huggingface/label-files' lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowerCamelCase__ : int = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : List[str] = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowerCamelCase__ : List[Any] = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowerCamelCase__ : Dict = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowerCamelCase__ : Optional[Any] = [2, 2, 20] lowerCamelCase__ : Optional[int] = [3, 12, 16] lowerCamelCase__ : str = [192, 768, 1024] lowerCamelCase__ : Any = CvtForImageClassification(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowerCamelCase__ : Optional[int] = OrderedDict() lowerCamelCase__ : Tuple = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowerCamelCase__ : Optional[Any] = list_of_state_dict + cls_token(_UpperCAmelCase ) lowerCamelCase__ : str = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowerCamelCase__ : str = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : int = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : str = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( """--cvt_model""", default="""cvt-w24""", type=str, help="""Name of the cvt model you'd like to convert.""", ) parser.add_argument( """--image_size""", default=3_84, type=int, help="""Input Image Size""", ) parser.add_argument( """--cvt_file_name""", default=R"""cvtmodels\CvT-w24-384x384-IN-22k.pth""", type=str, help="""Input Image Size""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
50
1
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Any = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=False ) -> List[str]: lowerCamelCase__ : Any = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('head' ): lowerCamelCase__ : Optional[int] = 'segformer.encoder.' + key if key.startswith('backbone' ): lowerCamelCase__ : Tuple = key.replace('backbone' , 'segformer.encoder' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : str = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Union[str, Any] = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : Any = key.replace('norm' , 'layer_norm' ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('segformer.encoder.layer_norm' ) + len('segformer.encoder.layer_norm' )] lowerCamelCase__ : Tuple = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[Any] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : List[str] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : Optional[int] = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Dict = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : str = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Any = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Tuple = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Optional[Any] = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Optional[Any] = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : Union[str, Any] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Tuple = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Tuple = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if key.startswith('head' ): lowerCamelCase__ : str = key.replace('head' , 'classifier' ) lowerCamelCase__ : Union[str, Any] = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : List[Any] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : List[Any] = state_dict.pop(F"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : int = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : List[str] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Union[str, Any] = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : str = kv_bias[ config.hidden_sizes[i] : ] def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: lowerCamelCase__ : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Any = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : List[str] = SegformerConfig() lowerCamelCase__ : Optional[int] = False # set attributes based on model_name lowerCamelCase__ : Optional[int] = 'huggingface/label-files' if "segformer" in model_name: lowerCamelCase__ : Optional[Any] = model_name[len('segformer.' ) : len('segformer.' ) + 2] if "ade" in model_name: lowerCamelCase__ : Optional[Any] = 150 lowerCamelCase__ : List[str] = 'ade20k-id2label.json' lowerCamelCase__ : int = (1, 150, 128, 128) elif "city" in model_name: lowerCamelCase__ : str = 19 lowerCamelCase__ : Tuple = 'cityscapes-id2label.json' lowerCamelCase__ : Optional[int] = (1, 19, 128, 128) else: raise ValueError(F"""Model {model_name} not supported""" ) elif "mit" in model_name: lowerCamelCase__ : int = True lowerCamelCase__ : Tuple = model_name[4:6] lowerCamelCase__ : Tuple = 1000 lowerCamelCase__ : Union[str, Any] = 'imagenet-1k-id2label.json' lowerCamelCase__ : int = (1, 1000) else: raise ValueError(F"""Model {model_name} not supported""" ) # set config attributes lowerCamelCase__ : Dict = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase__ : Optional[Any] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Union[str, Any] = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": lowerCamelCase__ : int = [64, 128, 320, 512] lowerCamelCase__ : Any = 256 elif size == "b2": lowerCamelCase__ : str = [64, 128, 320, 512] lowerCamelCase__ : Any = 768 lowerCamelCase__ : int = [3, 4, 6, 3] elif size == "b3": lowerCamelCase__ : List[str] = [64, 128, 320, 512] lowerCamelCase__ : Optional[Any] = 768 lowerCamelCase__ : List[Any] = [3, 4, 18, 3] elif size == "b4": lowerCamelCase__ : int = [64, 128, 320, 512] lowerCamelCase__ : Dict = 768 lowerCamelCase__ : List[str] = [3, 8, 27, 3] elif size == "b5": lowerCamelCase__ : Optional[Any] = [64, 128, 320, 512] lowerCamelCase__ : List[Any] = 768 lowerCamelCase__ : List[str] = [3, 6, 40, 3] else: raise ValueError(F"""Size {size} not supported""" ) # load image processor (only resize + normalize) lowerCamelCase__ : Tuple = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_UpperCAmelCase , align=_UpperCAmelCase , do_random_crop=_UpperCAmelCase ) # prepare image lowerCamelCase__ : List[str] = prepare_img() lowerCamelCase__ : int = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info(F"""Converting model {model_name}...""" ) # load original state dict if encoder_only: lowerCamelCase__ : int = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) else: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) )['state_dict'] # rename keys lowerCamelCase__ : Optional[Any] = rename_keys(_UpperCAmelCase , encoder_only=_UpperCAmelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict if encoder_only: lowerCamelCase__ : List[Any] = False lowerCamelCase__ : Optional[Any] = SegformerForImageClassification(_UpperCAmelCase ) else: lowerCamelCase__ : str = SegformerForSemanticSegmentation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : Union[str, Any] = model(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": lowerCamelCase__ : List[Any] = torch.tensor( [ [[-4.6_310, -5.5_232, -6.2_356], [-5.1_921, -6.1_444, -6.5_996], [-5.4_424, -6.2_790, -6.7_574]], [[-12.1_391, -13.3_122, -13.9_554], [-12.8_732, -13.9_352, -14.3_563], [-12.9_438, -13.8_226, -14.2_513]], [[-12.5_134, -13.4_686, -14.4_915], [-12.8_669, -14.4_343, -14.7_758], [-13.2_523, -14.5_819, -15.0_694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": lowerCamelCase__ : int = torch.tensor( [ [[-7.5_820, -8.7_231, -8.3_215], [-8.0_600, -10.3_529, -10.0_304], [-7.5_208, -9.4_103, -9.6_239]], [[-12.6_918, -13.8_994, -13.7_137], [-13.3_196, -15.7_523, -15.4_789], [-12.9_343, -14.8_757, -14.9_689]], [[-11.1_911, -11.9_421, -11.3_243], [-11.3_342, -13.6_839, -13.3_581], [-10.3_909, -12.1_832, -12.4_858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": lowerCamelCase__ : Tuple = torch.tensor( [ [[-11.8_173, -14.3_850, -16.3_128], [-14.5_648, -16.5_804, -18.6_568], [-14.7_223, -15.7_387, -18.4_218]], [[-15.7_290, -17.9_171, -19.4_423], [-18.3_105, -19.9_448, -21.4_661], [-17.9_296, -18.6_497, -20.7_910]], [[-15.0_783, -17.0_336, -18.2_789], [-16.8_771, -18.6_870, -20.1_612], [-16.2_454, -17.1_426, -19.5_055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": lowerCamelCase__ : str = torch.tensor( [ [[-9.0_878, -10.2_081, -10.1_891], [-9.3_144, -10.7_941, -10.9_843], [-9.2_294, -10.3_855, -10.5_704]], [[-12.2_316, -13.9_068, -13.6_102], [-12.9_161, -14.3_702, -14.3_235], [-12.5_233, -13.7_174, -13.7_932]], [[-14.6_275, -15.2_490, -14.9_727], [-14.3_400, -15.9_687, -16.2_827], [-14.1_484, -15.4_033, -15.8_937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": lowerCamelCase__ : int = torch.tensor( [ [[-12.3_144, -13.2_447, -14.0_802], [-13.3_614, -14.5_816, -15.6_117], [-13.3_340, -14.4_433, -16.2_219]], [[-19.2_781, -20.4_128, -20.7_506], [-20.6_153, -21.6_566, -22.0_998], [-19.9_800, -21.0_430, -22.1_494]], [[-18.8_739, -19.7_804, -21.1_834], [-20.1_233, -21.6_765, -23.2_944], [-20.0_315, -21.2_641, -23.6_944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": lowerCamelCase__ : Dict = torch.tensor( [ [[-9.5_524, -12.0_835, -11.7_348], [-10.5_229, -13.6_446, -14.5_662], [-9.5_842, -12.8_851, -13.9_414]], [[-15.3_432, -17.5_323, -17.0_818], [-16.3_330, -18.9_255, -19.2_101], [-15.1_340, -17.7_848, -18.3_971]], [[-12.6_072, -14.9_486, -14.6_631], [-13.7_629, -17.0_907, -17.7_745], [-12.7_899, -16.1_695, -17.1_671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": lowerCamelCase__ : int = torch.tensor( [ [[-11.9_295, -13.4_057, -14.8_106], [-13.3_431, -14.8_179, -15.3_781], [-14.2_836, -15.5_942, -16.1_588]], [[-11.4_906, -12.8_067, -13.6_564], [-13.1_189, -14.0_500, -14.1_543], [-13.8_748, -14.5_136, -14.8_789]], [[0.5_374, 0.1_067, -0.4_742], [0.1_141, -0.2_255, -0.7_099], [-0.3_000, -0.5_924, -1.3_105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": lowerCamelCase__ : int = torch.tensor( [ [[-7.8_217, -9.8_767, -10.1_717], [-9.4_438, -10.9_058, -11.4_047], [-9.7_939, -12.3_495, -12.1_079]], [[-7.1_514, -9.5_336, -10.0_860], [-9.7_776, -11.6_822, -11.8_439], [-10.1_411, -12.7_655, -12.8_972]], [[0.3_021, 0.0_805, -0.2_310], [-0.0_328, -0.1_605, -0.2_714], [-0.1_408, -0.5_477, -0.6_976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": lowerCamelCase__ : Optional[int] = torch.tensor( [ [ [-1.1_372e01, -1.2_787e01, -1.3_477e01], [-1.2_536e01, -1.4_194e01, -1.4_409e01], [-1.3_217e01, -1.4_888e01, -1.5_327e01], ], [ [-1.4_791e01, -1.7_122e01, -1.8_277e01], [-1.7_163e01, -1.9_192e01, -1.9_533e01], [-1.7_897e01, -1.9_991e01, -2.0_315e01], ], [ [7.6_723e-01, 4.1_921e-01, -7.7_878e-02], [4.7_772e-01, 9.5_557e-03, -2.8_082e-01], [3.6_032e-01, -2.4_826e-01, -5.1_168e-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": lowerCamelCase__ : Any = torch.tensor( [ [[-9.4_959, -11.3_087, -11.7_479], [-11.0_025, -12.6_540, -12.3_319], [-11.4_064, -13.0_487, -12.9_905]], [[-9.8_905, -11.3_084, -12.0_854], [-11.1_726, -12.7_698, -12.9_583], [-11.5_985, -13.3_278, -14.1_774]], [[0.2_213, 0.0_192, -0.2_466], [-0.1_731, -0.4_213, -0.4_874], [-0.3_126, -0.6_541, -1.1_389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": lowerCamelCase__ : Optional[int] = torch.tensor( [ [[-13.5_748, -13.9_111, -12.6_500], [-14.3_500, -15.3_683, -14.2_328], [-14.7_532, -16.0_424, -15.6_087]], [[-17.1_651, -15.8_725, -12.9_653], [-17.2_580, -17.3_718, -14.8_223], [-16.6_058, -16.8_783, -16.7_452]], [[-3.6_456, -3.0_209, -1.4_203], [-3.0_797, -3.1_959, -2.0_000], [-1.8_757, -1.9_217, -1.6_997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": lowerCamelCase__ : Union[str, Any] = torch.tensor( [ [[-16.0_976, -16.4_856, -17.3_962], [-16.6_234, -19.0_342, -19.7_685], [-16.0_900, -18.0_661, -19.1_180]], [[-18.4_750, -18.8_488, -19.5_074], [-19.4_030, -22.1_570, -22.5_977], [-19.1_191, -20.8_486, -22.3_783]], [[-4.5_178, -5.5_037, -6.5_109], [-5.0_884, -7.2_174, -8.0_334], [-4.4_156, -5.8_117, -7.2_970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": lowerCamelCase__ : Union[str, Any] = torch.tensor( [ [[-14.2_081, -14.4_732, -14.1_977], [-14.5_867, -16.4_423, -16.6_356], [-13.4_441, -14.9_685, -16.8_696]], [[-14.4_576, -14.7_073, -15.0_451], [-15.0_816, -17.6_237, -17.9_873], [-14.4_213, -16.0_199, -18.5_992]], [[-4.7_349, -4.9_588, -5.0_966], [-4.3_210, -6.9_325, -7.2_591], [-3.4_312, -4.7_484, -7.1_917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": lowerCamelCase__ : Optional[Any] = torch.tensor( [ [[-11.7_737, -11.9_526, -11.3_273], [-13.6_692, -14.4_574, -13.8_878], [-13.8_937, -14.6_924, -15.9_345]], [[-14.6_706, -14.5_330, -14.1_306], [-16.1_502, -16.8_180, -16.4_269], [-16.8_338, -17.8_939, -20.1_746]], [[1.0_491, 0.8_289, 1.0_310], [1.1_044, 0.5_219, 0.8_055], [1.0_899, 0.6_926, 0.5_590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": lowerCamelCase__ : Any = torch.tensor( [ [[-12.5_641, -13.4_777, -13.0_684], [-13.9_587, -15.8_983, -16.6_557], [-13.3_109, -15.7_350, -16.3_141]], [[-14.7_074, -15.4_352, -14.5_944], [-16.6_353, -18.1_663, -18.6_120], [-15.1_702, -18.0_329, -18.1_547]], [[-1.7_990, -2.0_951, -1.7_784], [-2.6_397, -3.8_245, -3.9_686], [-1.5_264, -2.8_126, -2.9_316]], ] ) else: lowerCamelCase__ : Union[str, Any] = logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , _UpperCAmelCase , atol=1e-2 ) # finally, save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( """--model_name""", default="""segformer.b0.512x512.ade.160k""", type=str, help="""Name of the model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) _UpperCAmelCase : str = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
50
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : str = { """configuration_time_series_transformer""": [ """TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimeSeriesTransformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ """TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimeSeriesTransformerForPrediction""", """TimeSeriesTransformerModel""", """TimeSeriesTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
1
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: if "cls_token" in name: lowerCamelCase__ : List[str] = name.replace('cls_token' , 'vit.embeddings.cls_token' ) if "mask_token" in name: lowerCamelCase__ : Optional[int] = name.replace('mask_token' , 'decoder.mask_token' ) if "decoder_pos_embed" in name: lowerCamelCase__ : Optional[Any] = name.replace('decoder_pos_embed' , 'decoder.decoder_pos_embed' ) if "pos_embed" in name and "decoder" not in name: lowerCamelCase__ : Dict = name.replace('pos_embed' , 'vit.embeddings.position_embeddings' ) if "patch_embed.proj" in name: lowerCamelCase__ : Union[str, Any] = name.replace('patch_embed.proj' , 'vit.embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: lowerCamelCase__ : int = name.replace('patch_embed.norm' , 'vit.embeddings.norm' ) if "decoder_blocks" in name: lowerCamelCase__ : Optional[int] = name.replace('decoder_blocks' , 'decoder.decoder_layers' ) if "blocks" in name: lowerCamelCase__ : str = name.replace('blocks' , 'vit.encoder.layer' ) if "attn.proj" in name: lowerCamelCase__ : Dict = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: lowerCamelCase__ : int = name.replace('attn' , 'attention.self' ) if "norm1" in name: lowerCamelCase__ : Any = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: lowerCamelCase__ : int = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: lowerCamelCase__ : str = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: lowerCamelCase__ : int = name.replace('mlp.fc2' , 'output.dense' ) if "decoder_embed" in name: lowerCamelCase__ : Dict = name.replace('decoder_embed' , 'decoder.decoder_embed' ) if "decoder_norm" in name: lowerCamelCase__ : Any = name.replace('decoder_norm' , 'decoder.decoder_norm' ) if "decoder_pred" in name: lowerCamelCase__ : List[Any] = name.replace('decoder_pred' , 'decoder.decoder_pred' ) if "norm.weight" in name and "decoder" not in name: lowerCamelCase__ : str = name.replace('norm.weight' , 'vit.layernorm.weight' ) if "norm.bias" in name and "decoder" not in name: lowerCamelCase__ : List[Any] = name.replace('norm.bias' , 'vit.layernorm.bias' ) return name def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Tuple = orig_state_dict.pop(_UpperCAmelCase ) if "qkv" in key: lowerCamelCase__ : List[str] = key.split('.' ) lowerCamelCase__ : List[str] = int(key_split[1] ) if "decoder_blocks" in key: lowerCamelCase__ : Any = config.decoder_hidden_size lowerCamelCase__ : List[Any] = 'decoder.decoder_layers.' if "weight" in key: lowerCamelCase__ : int = val[:dim, :] lowerCamelCase__ : Union[str, Any] = val[dim : dim * 2, :] lowerCamelCase__ : Union[str, Any] = val[-dim:, :] elif "bias" in key: lowerCamelCase__ : List[str] = val[:dim] lowerCamelCase__ : Optional[int] = val[dim : dim * 2] lowerCamelCase__ : str = val[-dim:] else: lowerCamelCase__ : Any = config.hidden_size lowerCamelCase__ : Dict = 'vit.encoder.layer.' if "weight" in key: lowerCamelCase__ : Optional[Any] = val[:dim, :] lowerCamelCase__ : List[str] = val[dim : dim * 2, :] lowerCamelCase__ : List[str] = val[-dim:, :] elif "bias" in key: lowerCamelCase__ : Dict = val[:dim] lowerCamelCase__ : Union[str, Any] = val[dim : dim * 2] lowerCamelCase__ : Optional[Any] = val[-dim:] else: lowerCamelCase__ : Dict = val return orig_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ : Union[str, Any] = ViTMAEConfig() if "large" in checkpoint_url: lowerCamelCase__ : List[Any] = 1024 lowerCamelCase__ : str = 4096 lowerCamelCase__ : Optional[Any] = 24 lowerCamelCase__ : List[str] = 16 elif "huge" in checkpoint_url: lowerCamelCase__ : Any = 14 lowerCamelCase__ : str = 1280 lowerCamelCase__ : Dict = 5120 lowerCamelCase__ : Tuple = 32 lowerCamelCase__ : List[Any] = 16 lowerCamelCase__ : List[Any] = ViTMAEForPreTraining(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.hub.load_state_dict_from_url(_UpperCAmelCase , map_location='cpu' )['model'] lowerCamelCase__ : List[str] = ViTMAEImageProcessor(size=config.image_size ) lowerCamelCase__ : int = convert_state_dict(_UpperCAmelCase , _UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() lowerCamelCase__ : Dict = 'https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg' lowerCamelCase__ : Union[str, Any] = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) lowerCamelCase__ : int = ViTMAEImageProcessor(size=config.image_size ) lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ) # forward pass torch.manual_seed(2 ) lowerCamelCase__ : str = model(**_UpperCAmelCase ) lowerCamelCase__ : Any = outputs.logits if "large" in checkpoint_url: lowerCamelCase__ : List[str] = torch.tensor( [[-0.7_309, -0.7_128, -1.0_169], [-1.0_161, -0.9_058, -1.1_878], [-1.0_478, -0.9_411, -1.1_911]] ) elif "huge" in checkpoint_url: lowerCamelCase__ : Optional[Any] = torch.tensor( [[-1.1_599, -0.9_199, -1.2_221], [-1.1_952, -0.9_269, -1.2_307], [-1.2_143, -0.9_337, -1.2_262]] ) else: lowerCamelCase__ : Tuple = torch.tensor( [[-0.9_192, -0.8_481, -1.1_259], [-1.1_349, -1.0_034, -1.2_599], [-1.1_757, -1.0_429, -1.2_726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth""", type=str, help="""URL of the checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
50
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : int = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : int = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } _UpperCAmelCase : Any = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Tuple=None , UpperCAmelCase : Any=None , UpperCAmelCase : List[Any]=True , UpperCAmelCase : Optional[Any]="[UNK]" , UpperCAmelCase : Any="[SEP]" , UpperCAmelCase : Tuple="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Any=None , **UpperCAmelCase : Optional[int] , ) -> str: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Optional[int] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Optional[Any] = do_lower_case lowerCamelCase__ : str = strip_accents lowerCamelCase__ : Optional[Any] = tokenize_chinese_chars lowerCamelCase__ : int = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : str = do_lower_case def A_ ( self : Optional[int] , UpperCAmelCase : int , **UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH lowerCamelCase__ : Optional[int] = text lowerCamelCase__ : Dict = kwargs.pop('text_pair' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = kwargs.pop('return_tensors' , UpperCAmelCase ) lowerCamelCase__ : List[Any] = { 'input_ids': [], 'attention_mask': [], 'token_type_ids': [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: lowerCamelCase__ : Tuple = batch_text_pair[idx] else: lowerCamelCase__ : Dict = None lowerCamelCase__ : Optional[int] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Any = encoded_candidates.get('input_ids' ) lowerCamelCase__ : Union[str, Any] = encoded_candidates.get('attention_mask' ) lowerCamelCase__ : Tuple = encoded_candidates.get('token_type_ids' ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) lowerCamelCase__ : int = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : int , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ) -> List[str]: lowerCamelCase__ : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Tuple , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : int = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase : Optional[Any] = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[Any] = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys _UpperCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _UpperCAmelCase : Optional[Any] = logging.getLogger(__name__) @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None @dataclass(frozen=__UpperCamelCase ) class lowerCAmelCase : UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if is_torch_available(): import torch from torch.utils.data import Dataset class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 def __init__( self : int , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : List[str]=False , UpperCAmelCase : bool = False , ) -> List[str]: lowerCamelCase__ : int = hans_processors[task]() lowerCamelCase__ : Optional[Any] = os.path.join( UpperCAmelCase , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(UpperCAmelCase ) , UpperCAmelCase , ) , ) lowerCamelCase__ : int = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = label_list[2], label_list[1] lowerCamelCase__ : List[str] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : str = cached_features_file + '.lock' with FileLock(UpperCAmelCase ): if os.path.exists(UpperCAmelCase ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) lowerCamelCase__ : int = torch.load(UpperCAmelCase ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) lowerCamelCase__ : str = ( processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) ) logger.info('Training examples: %s' , len(UpperCAmelCase ) ) lowerCamelCase__ : Dict = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) logger.info('Saving features into cached file %s' , UpperCAmelCase ) torch.save(self.features , UpperCAmelCase ) def __len__( self : Optional[int] ) -> Optional[Any]: return len(self.features ) def __getitem__( self : Tuple , UpperCAmelCase : Dict ) -> InputFeatures: return self.features[i] def A_ ( self : int ) -> int: return self.label_list if is_tf_available(): import tensorflow as tf class lowerCAmelCase : UpperCAmelCase__ = 42 def __init__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : str , UpperCAmelCase : Optional[int] = 128 , UpperCAmelCase : Any=False , UpperCAmelCase : bool = False , ) -> Union[str, Any]: lowerCamelCase__ : Any = hans_processors[task]() lowerCamelCase__ : Optional[Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : str = label_list[2], label_list[1] lowerCamelCase__ : Optional[int] = label_list lowerCamelCase__ : int = processor.get_dev_examples(UpperCAmelCase ) if evaluate else processor.get_train_examples(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = hans_convert_examples_to_features(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(UpperCAmelCase )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCamelCase__ : Optional[int] = tf.data.Dataset.from_generator( UpperCAmelCase , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def A_ ( self : Any ) -> Any: return self.dataset def __len__( self : Tuple ) -> int: return len(self.features ) def __getitem__( self : List[str] , UpperCAmelCase : Any ) -> InputFeatures: return self.features[i] def A_ ( self : Dict ) -> str: return self.label_list class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : int , UpperCAmelCase : List[Any] ) -> int: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_train_set.txt' ) ) , 'train' ) def A_ ( self : Any , UpperCAmelCase : int ) -> List[Any]: return self._create_examples(self._read_tsv(os.path.join(UpperCAmelCase , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def A_ ( self : Any ) -> List[Any]: return ["contradiction", "entailment", "neutral"] def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : List[str] ) -> List[str]: lowerCamelCase__ : List[str] = [] for i, line in enumerate(UpperCAmelCase ): if i == 0: continue lowerCamelCase__ : Tuple = '%s-%s' % (set_type, line[0]) lowerCamelCase__ : str = line[5] lowerCamelCase__ : Dict = line[6] lowerCamelCase__ : int = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCamelCase__ : Dict = line[0] examples.append(InputExample(guid=UpperCAmelCase , text_a=UpperCAmelCase , text_b=UpperCAmelCase , label=UpperCAmelCase , pairID=UpperCAmelCase ) ) return examples def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> Optional[int]: lowerCamelCase__ : int = {label: i for i, label in enumerate(_UpperCAmelCase )} lowerCamelCase__ : List[Any] = [] for ex_index, example in tqdm.tqdm(enumerate(_UpperCAmelCase ) , desc='convert examples to features' ): if ex_index % 1_0000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCamelCase__ : List[Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=_UpperCAmelCase , max_length=_UpperCAmelCase , padding='max_length' , truncation=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , ) lowerCamelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCamelCase__ : Optional[int] = int(example.pairID ) features.append(InputFeatures(**_UpperCAmelCase , label=_UpperCAmelCase , pairID=_UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F"""guid: {example}""" ) logger.info(F"""features: {features[i]}""" ) return features _UpperCAmelCase : str = { """hans""": 3, } _UpperCAmelCase : List[Any] = { """hans""": HansProcessor, }
50
1
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: lowerCamelCase__ : Union[str, Any] = 2 lowerCamelCase__ : Tuple = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(_UpperCAmelCase ) if n > 1: factors.append(_UpperCAmelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
50
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCAmelCase : Optional[Any] = """▁""" _UpperCAmelCase : Optional[Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = BertGenerationTokenizer UpperCAmelCase__ = False UpperCAmelCase__ = True def A_ ( self : List[Any] ) -> List[str]: super().setUp() lowerCamelCase__ : Dict = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : List[str] = '<s>' lowerCamelCase__ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : List[str] ) -> Optional[int]: lowerCamelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(UpperCAmelCase ) , 1002 ) def A_ ( self : List[Any] ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = BertGenerationTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.tokenize('This is a test' ) self.assertListEqual(UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowerCamelCase__ : List[str] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase__ : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowerCamelCase__ : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def A_ ( self : Dict ) -> Tuple: return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def A_ ( self : Optional[int] ) -> List[str]: lowerCamelCase__ : Union[str, Any] = 'Hello World!' lowerCamelCase__ : Dict = [18536, 2260, 101] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @slow def A_ ( self : Optional[Any] ) -> str: lowerCamelCase__ : List[Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowerCamelCase__ : Any = [ 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, ] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @require_torch @slow def A_ ( self : int ) -> Optional[Any]: import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence lowerCamelCase__ : str = list(self.big_tokenizer.get_vocab().keys() )[:10] lowerCamelCase__ : int = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.big_tokenizer.encode_plus(UpperCAmelCase , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=UpperCAmelCase ) lowerCamelCase__ : Tuple = BertGenerationConfig() lowerCamelCase__ : Optional[Any] = BertGenerationEncoder(UpperCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase ) model(**UpperCAmelCase ) @slow def A_ ( self : Optional[int] ) -> List[Any]: # fmt: off lowerCamelCase__ : Any = {'input_ids': [[39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114], [448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
50
1
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = KandinskyVaaControlnetPipeline UpperCAmelCase__ = ["""image_embeds""", """negative_image_embeds""", """hint"""] UpperCAmelCase__ = ["""image_embeds""", """negative_image_embeds""", """hint"""] UpperCAmelCase__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] UpperCAmelCase__ = False @property def A_ ( self : Union[str, Any] ) -> List[Any]: return 32 @property def A_ ( self : Optional[int] ) -> str: return 32 @property def A_ ( self : str ) -> Tuple: return self.time_input_dim @property def A_ ( self : int ) -> Any: return self.time_input_dim * 4 @property def A_ ( self : List[str] ) -> int: return 100 @property def A_ ( self : int ) -> Union[str, Any]: torch.manual_seed(0 ) lowerCamelCase__ : List[str] = { 'in_channels': 8, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image_hint', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } lowerCamelCase__ : Dict = UNetaDConditionModel(**UpperCAmelCase ) return model @property def A_ ( self : Any ) -> Optional[int]: return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def A_ ( self : Union[str, Any] ) -> List[str]: torch.manual_seed(0 ) lowerCamelCase__ : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def A_ ( self : Tuple ) -> Dict: lowerCamelCase__ : Optional[Any] = self.dummy_unet lowerCamelCase__ : str = self.dummy_movq lowerCamelCase__ : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule='linear' , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=UpperCAmelCase , set_alpha_to_one=UpperCAmelCase , steps_offset=1 , prediction_type='epsilon' , thresholding=UpperCAmelCase , ) lowerCamelCase__ : Optional[Any] = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def A_ ( self : str , UpperCAmelCase : Tuple , UpperCAmelCase : str=0 ) -> List[Any]: lowerCamelCase__ : Any = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(UpperCAmelCase ) ).to(UpperCAmelCase ) lowerCamelCase__ : Dict = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( UpperCAmelCase ) # create hint lowerCamelCase__ : Any = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCAmelCase ) ).to(UpperCAmelCase ) if str(UpperCAmelCase ).startswith('mps' ): lowerCamelCase__ : Dict = torch.manual_seed(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = torch.Generator(device=UpperCAmelCase ).manual_seed(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = { 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'hint': hint, 'generator': generator, 'height': 64, 'width': 64, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def A_ ( self : List[Any] ) -> Any: lowerCamelCase__ : int = 'cpu' lowerCamelCase__ : List[str] = self.get_dummy_components() lowerCamelCase__ : List[str] = self.pipeline_class(**UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = pipe.to(UpperCAmelCase ) pipe.set_progress_bar_config(disable=UpperCAmelCase ) lowerCamelCase__ : Dict = pipe(**self.get_dummy_inputs(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = output.images lowerCamelCase__ : Union[str, Any] = pipe( **self.get_dummy_inputs(UpperCAmelCase ) , return_dict=UpperCAmelCase , )[0] lowerCamelCase__ : List[Any] = image[0, -3:, -3:, -1] lowerCamelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase__ : Dict = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCAmelCase ( unittest.TestCase ): def A_ ( self : Optional[int] ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self : Optional[int] ) -> Optional[Any]: lowerCamelCase__ : int = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy' ) lowerCamelCase__ : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/hint_image_cat.png' ) lowerCamelCase__ : str = torch.from_numpy(np.array(UpperCAmelCase ) ).float() / 2_5_5.0 lowerCamelCase__ : Optional[Any] = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowerCamelCase__ : Optional[Any] = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = KandinskyVaaControlnetPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-controlnet-depth' , torch_dtype=torch.floataa ) lowerCamelCase__ : str = pipeline.to(UpperCAmelCase ) pipeline.set_progress_bar_config(disable=UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = 'A robot, 4k photo' lowerCamelCase__ : int = torch.Generator(device='cuda' ).manual_seed(0 ) lowerCamelCase__ , lowerCamelCase__ : Tuple = pipe_prior( UpperCAmelCase , generator=UpperCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() lowerCamelCase__ : Union[str, Any] = torch.Generator(device='cuda' ).manual_seed(0 ) lowerCamelCase__ : List[Any] = pipeline( image_embeds=UpperCAmelCase , negative_image_embeds=UpperCAmelCase , hint=UpperCAmelCase , generator=UpperCAmelCase , num_inference_steps=100 , output_type='np' , ) lowerCamelCase__ : List[str] = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(UpperCAmelCase , UpperCAmelCase )
50
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _UpperCAmelCase : str = pytest.mark.integration @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : List[Any] ) -> Union[str, Any]: lowerCamelCase__ : List[Any] = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(UpperCAmelCase ) for x in np.arange(30 ).tolist()]} ) return dset def A_ ( self : Optional[Any] ) -> Optional[int]: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() lowerCamelCase__ : List[Any] = dset.map( lambda UpperCAmelCase , UpperCAmelCase : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=UpperCAmelCase , keep_in_memory=UpperCAmelCase ) lowerCamelCase__ : Tuple = dset.add_faiss_index('vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def A_ ( self : Union[str, Any] ) -> int: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : List[str] ) -> Tuple: import faiss lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ , lowerCamelCase__ : str = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def A_ ( self : Any ) -> Optional[Any]: lowerCamelCase__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(UpperCAmelCase , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def A_ ( self : Dict ) -> Dict: from elasticsearch import Elasticsearch lowerCamelCase__ : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : List[Any] = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) lowerCamelCase__ : int = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} lowerCamelCase__ : List[str] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Dict = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Any ) -> Dict: import faiss lowerCamelCase__ : Tuple = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query lowerCamelCase__ : int = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Any = 1 lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = index.search(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries lowerCamelCase__ : str = np.eye(5 , dtype=np.floataa )[::-1] lowerCamelCase__ , lowerCamelCase__ : Dict = index.search_batch(UpperCAmelCase ) self.assertRaises(UpperCAmelCase , index.search_batch , queries[0] ) lowerCamelCase__ : str = [scores[0] for scores in total_scores] lowerCamelCase__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , UpperCAmelCase ) def A_ ( self : List[Any] ) -> List[Any]: import faiss lowerCamelCase__ : Any = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) lowerCamelCase__ : Tuple = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(UpperCAmelCase ): lowerCamelCase__ : List[str] = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def A_ ( self : List[str] ) -> Optional[int]: import faiss lowerCamelCase__ : Optional[Any] = faiss.IndexFlat(5 ) lowerCamelCase__ : int = FaissIndex(custom_index=UpperCAmelCase ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def A_ ( self : Any ) -> Optional[int]: import faiss lowerCamelCase__ : Any = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCAmelCase ) as tmp_file: index.save(tmp_file.name ) lowerCamelCase__ : List[Any] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) lowerCamelCase__ : List[str] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Tuple = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(UpperCAmelCase ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: import faiss lowerCamelCase__ : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) lowerCamelCase__ : Optional[int] = 'index.faiss' lowerCamelCase__ : Optional[Any] = F"""mock://{index_name}""" index.save(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Tuple = FaissIndex.load(_UpperCAmelCase , storage_options=mockfs.storage_options ) lowerCamelCase__ : Optional[int] = np.zeros(5 , dtype=np.floataa ) lowerCamelCase__ : Dict = 1 lowerCamelCase__ , lowerCamelCase__ : str = index.search(_UpperCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Dict ) -> List[Any]: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: lowerCamelCase__ : Any = Elasticsearch() lowerCamelCase__ : Tuple = {'acknowledged': True} lowerCamelCase__ : Tuple = ElasticSearchIndex(es_client=UpperCAmelCase ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query lowerCamelCase__ : Optional[int] = 'foo' lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search(UpperCAmelCase ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout lowerCamelCase__ : Any = 'foo' lowerCamelCase__ : List[str] = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} lowerCamelCase__ , lowerCamelCase__ : Tuple = index.search(UpperCAmelCase , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries lowerCamelCase__ : List[str] = ['foo', 'bar', 'foobar'] lowerCamelCase__ : str = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : str = index.search_batch(UpperCAmelCase ) lowerCamelCase__ : List[str] = [scores[0] for scores in total_scores] lowerCamelCase__ : List[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase ) # batched queries with timeout lowerCamelCase__ : str = ['foo', 'bar', 'foobar'] lowerCamelCase__ : List[Any] = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = index.search_batch(UpperCAmelCase , request_timeout=30 ) lowerCamelCase__ : Optional[Any] = [scores[0] for scores in total_scores] lowerCamelCase__ : Dict = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCAmelCase ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCAmelCase )
50
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowerCAmelCase ( __UpperCamelCase ): def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCAmelCase , 'tf_padding' ) ) self.parent.assertTrue(hasattr(UpperCAmelCase , 'depth_multiplier' ) ) class lowerCAmelCase : def __init__( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[Any]=13 , UpperCAmelCase : str=3 , UpperCAmelCase : Optional[int]=32 , UpperCAmelCase : Union[str, Any]=0.2_5 , UpperCAmelCase : Union[str, Any]=8 , UpperCAmelCase : List[Any]=8 , UpperCAmelCase : List[str]=6 , UpperCAmelCase : Dict=32 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : List[str]=True , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : List[str]="relu6" , UpperCAmelCase : Dict=1280 , UpperCAmelCase : int=0.1 , UpperCAmelCase : Optional[Any]=0.0_2 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : Dict=True , UpperCAmelCase : int=10 , UpperCAmelCase : str=None , ) -> Optional[Any]: lowerCamelCase__ : Tuple = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[Any] = image_size lowerCamelCase__ : List[Any] = depth_multiplier lowerCamelCase__ : List[Any] = depth_divisible_by lowerCamelCase__ : Dict = min_depth lowerCamelCase__ : Optional[Any] = expand_ratio lowerCamelCase__ : Union[str, Any] = tf_padding lowerCamelCase__ : Any = output_stride lowerCamelCase__ : List[str] = first_layer_is_expansion lowerCamelCase__ : Union[str, Any] = finegrained_output lowerCamelCase__ : Union[str, Any] = hidden_act lowerCamelCase__ : Any = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Union[str, Any] = classifier_dropout_prob lowerCamelCase__ : Tuple = use_labels lowerCamelCase__ : int = is_training lowerCamelCase__ : List[str] = num_labels lowerCamelCase__ : str = initializer_range lowerCamelCase__ : Optional[Any] = scope def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Tuple = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : List[str] = self.get_config() return config, pixel_values, labels, pixel_labels def A_ ( self : Any ) -> Union[str, Any]: return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def A_ ( self : Dict , UpperCAmelCase : str , UpperCAmelCase : List[Any] , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int] ) -> Any: lowerCamelCase__ : int = MobileNetVaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model(UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def A_ ( self : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Dict , UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: lowerCamelCase__ : Optional[int] = self.num_labels lowerCamelCase__ : Any = MobileNetVaForImageClassification(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : int = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A_ ( self : str , UpperCAmelCase : List[Any] , UpperCAmelCase : Optional[int] , UpperCAmelCase : int , UpperCAmelCase : int ) -> Tuple: lowerCamelCase__ : Any = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForSemanticSegmentation(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : str = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowerCamelCase__ : Optional[int] = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def A_ ( self : Optional[int] ) -> Union[str, Any]: lowerCamelCase__ : Dict = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Dict = config_and_inputs lowerCamelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) UpperCAmelCase__ = ( { """feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification, """image-segmentation""": MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : Tuple ) -> int: lowerCamelCase__ : Optional[Any] = MobileNetVaModelTester(self ) lowerCamelCase__ : int = MobileNetVaConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def A_ ( self : List[str] ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def A_ ( self : List[Any] ) -> List[Any]: pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def A_ ( self : Any ) -> Optional[Any]: pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def A_ ( self : str ) -> Optional[Any]: pass def A_ ( self : Union[str, Any] ) -> Tuple: lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(UpperCAmelCase ) lowerCamelCase__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : List[str] = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def A_ ( self : str ) -> Tuple: lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def A_ ( self : str ) -> Optional[int]: def check_hidden_states_output(UpperCAmelCase : List[str] , UpperCAmelCase : Any , UpperCAmelCase : List[Any] ): lowerCamelCase__ : Any = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Any = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = outputs.hidden_states lowerCamelCase__ : Dict = 16 self.assertEqual(len(UpperCAmelCase ) , UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> str: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) def A_ ( self : Optional[int] ) -> int: lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @slow def A_ ( self : Optional[Any] ) -> Tuple: for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Tuple = MobileNetVaModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowerCAmelCase ( unittest.TestCase ): @cached_property def A_ ( self : Dict ) -> List[str]: return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def A_ ( self : Any ) -> Union[str, Any]: lowerCamelCase__ : List[str] = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(UpperCAmelCase ) lowerCamelCase__ : Any = self.default_image_processor lowerCamelCase__ : Any = prepare_img() lowerCamelCase__ : Dict = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Dict = model(**UpperCAmelCase ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) lowerCamelCase__ : Dict = torch.tensor([0.2_4_4_5, -1.1_9_9_3, 0.1_9_0_5] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase , atol=1e-4 ) ) @slow def A_ ( self : Union[str, Any] ) -> Optional[Any]: lowerCamelCase__ : str = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) lowerCamelCase__ : List[str] = model.to(UpperCAmelCase ) lowerCamelCase__ : Dict = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) lowerCamelCase__ : Any = prepare_img() lowerCamelCase__ : Optional[int] = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**UpperCAmelCase ) lowerCamelCase__ : str = outputs.logits # verify the logits lowerCamelCase__ : Optional[Any] = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = torch.tensor( [ [[1_7.5_7_9_0, 1_7.7_5_8_1, 1_8.3_3_5_5], [1_8.3_2_5_7, 1_8.4_2_3_0, 1_8.8_9_7_3], [1_8.6_1_6_9, 1_8.8_6_5_0, 1_9.2_1_8_7]], [[-2.1_5_9_5, -2.0_9_7_7, -2.3_7_4_1], [-2.4_2_2_6, -2.3_0_2_8, -2.6_8_3_5], [-2.7_8_1_9, -2.5_9_9_1, -2.7_7_0_6]], [[4.2_0_5_8, 4.8_3_1_7, 4.7_6_3_8], [4.4_1_3_6, 5.0_3_6_1, 4.9_3_8_3], [4.5_0_2_8, 4.9_6_4_4, 4.8_7_3_4]], ] , device=UpperCAmelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : str = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCamelCase__ : Tuple = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCamelCase__ : Dict = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCamelCase__ : str = subset[i - 1][j] if arr[i - 1] <= j: lowerCamelCase__ : Dict = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
50
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class lowerCAmelCase : UpperCAmelCase__ = MBartConfig UpperCAmelCase__ = {} UpperCAmelCase__ = """gelu""" def __init__( self : List[str] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : str=13 , UpperCAmelCase : Union[str, Any]=7 , UpperCAmelCase : Dict=True , UpperCAmelCase : Tuple=False , UpperCAmelCase : Dict=99 , UpperCAmelCase : int=32 , UpperCAmelCase : Tuple=2 , UpperCAmelCase : str=4 , UpperCAmelCase : Any=37 , UpperCAmelCase : int=0.1 , UpperCAmelCase : Union[str, Any]=0.1 , UpperCAmelCase : List[str]=20 , UpperCAmelCase : Dict=2 , UpperCAmelCase : Optional[int]=1 , UpperCAmelCase : Dict=0 , ) -> Dict: lowerCamelCase__ : Union[str, Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : str = seq_length lowerCamelCase__ : str = is_training lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : List[Any] = vocab_size lowerCamelCase__ : int = hidden_size lowerCamelCase__ : List[Any] = num_hidden_layers lowerCamelCase__ : Dict = num_attention_heads lowerCamelCase__ : Optional[Any] = intermediate_size lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase__ : Dict = max_position_embeddings lowerCamelCase__ : Dict = eos_token_id lowerCamelCase__ : Any = pad_token_id lowerCamelCase__ : int = bos_token_id def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCamelCase__ : Optional[int] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCamelCase__ : List[str] = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCamelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase__ : List[Any] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowerCamelCase__ : int = prepare_mbart_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, inputs_dict def A_ ( self : Optional[Any] , UpperCAmelCase : Optional[int] , UpperCAmelCase : Optional[int] ) -> Union[str, Any]: lowerCamelCase__ : Tuple = TFMBartModel(config=UpperCAmelCase ).get_decoder() lowerCamelCase__ : int = inputs_dict['input_ids'] lowerCamelCase__ : Union[str, Any] = input_ids[:1, :] lowerCamelCase__ : str = inputs_dict['attention_mask'][:1, :] lowerCamelCase__ : Union[str, Any] = inputs_dict['head_mask'] lowerCamelCase__ : Union[str, Any] = 1 # first forward pass lowerCamelCase__ : Union[str, Any] = model(UpperCAmelCase , attention_mask=UpperCAmelCase , head_mask=UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Any = outputs.to_tuple() lowerCamelCase__ : Optional[int] = past_key_values[1] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> Union[str, Any]: if attention_mask is None: lowerCamelCase__ : Union[str, Any] = tf.cast(tf.math.not_equal(_UpperCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCamelCase__ : Any = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCamelCase__ : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCamelCase__ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () UpperCAmelCase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase__ = ( { """conversational""": TFMBartForConditionalGeneration, """feature-extraction""": TFMBartModel, """summarization""": TFMBartForConditionalGeneration, """text2text-generation""": TFMBartForConditionalGeneration, """translation""": TFMBartForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : Tuple , UpperCAmelCase : Any , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Tuple , UpperCAmelCase : Tuple , UpperCAmelCase : str ) -> Tuple: if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def A_ ( self : Dict ) -> int: lowerCamelCase__ : Dict = TFMBartModelTester(self ) lowerCamelCase__ : List[str] = ConfigTester(self , config_class=UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: self.config_tester.run_common_tests() def A_ ( self : Tuple ) -> int: lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*UpperCAmelCase ) @require_sentencepiece @require_tokenizers @require_tf class lowerCAmelCase ( unittest.TestCase ): UpperCAmelCase__ = [ """ UN Chief Says There Is No Military Solution in Syria""", ] UpperCAmelCase__ = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", ] UpperCAmelCase__ = """facebook/mbart-large-en-ro""" @cached_property def A_ ( self : Union[str, Any] ) -> Optional[Any]: return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def A_ ( self : Optional[Any] ) -> Any: lowerCamelCase__ : str = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def A_ ( self : List[Any] , **UpperCAmelCase : Optional[Any] ) -> Optional[Any]: lowerCamelCase__ : List[str] = self.translate_src_text(**UpperCAmelCase ) self.assertListEqual(self.expected_text , UpperCAmelCase ) def A_ ( self : str , **UpperCAmelCase : int ) -> List[str]: lowerCamelCase__ : Union[str, Any] = self.tokenizer(self.src_text , **UpperCAmelCase , return_tensors='tf' ) lowerCamelCase__ : Optional[int] = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) lowerCamelCase__ : str = self.tokenizer.batch_decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase ) return generated_words @slow def A_ ( self : str ) -> Union[str, Any]: self._assert_generated_batch_equal_expected()
50
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """M-CLIP""" def __init__( self : Optional[Any] , UpperCAmelCase : Union[str, Any]=1024 , UpperCAmelCase : Tuple=768 , **UpperCAmelCase : Optional[int] ) -> Dict: lowerCamelCase__ : Optional[int] = transformerDimSize lowerCamelCase__ : Optional[Any] = imageDimSize super().__init__(**UpperCAmelCase ) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = MCLIPConfig def __init__( self : List[Any] , UpperCAmelCase : Dict , *UpperCAmelCase : Any , **UpperCAmelCase : Dict ) -> Dict: super().__init__(UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Tuple = XLMRobertaModel(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Dict ) -> Tuple: lowerCamelCase__ : Any = self.transformer(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase )[0] lowerCamelCase__ : int = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(UpperCAmelCase ), embs
50
1
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) @dataclass class lowerCAmelCase : UpperCAmelCase__ = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(glue_processors.keys() )} ) UpperCAmelCase__ = field( metadata={"""help""": """The input data dir. Should contain the .tsv files (or other data files) for the task."""} ) UpperCAmelCase__ = field( default=1_28, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) UpperCAmelCase__ = field( default=__UpperCamelCase, metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def A_ ( self : Optional[Any] ) -> Dict: lowerCamelCase__ : Optional[Any] = self.task_name.lower() class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """train""" UpperCAmelCase__ = """dev""" UpperCAmelCase__ = """test""" class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 def __init__( self : List[Any] , UpperCAmelCase : GlueDataTrainingArguments , UpperCAmelCase : PreTrainedTokenizerBase , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Union[str, Split] = Split.train , UpperCAmelCase : Optional[str] = None , ) -> int: warnings.warn( 'This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' , UpperCAmelCase , ) lowerCamelCase__ : List[Any] = args lowerCamelCase__ : List[Any] = glue_processors[args.task_name]() lowerCamelCase__ : Union[str, Any] = glue_output_modes[args.task_name] if isinstance(UpperCAmelCase , UpperCAmelCase ): try: lowerCamelCase__ : Union[str, Any] = Split[mode] except KeyError: raise KeyError('mode is not a valid split name' ) # Load data features from cache or dataset file lowerCamelCase__ : int = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}""" , ) lowerCamelCase__ : List[str] = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCamelCase__ , lowerCamelCase__ : Tuple = label_list[2], label_list[1] lowerCamelCase__ : int = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCamelCase__ : Union[str, Any] = cached_features_file + '.lock' with FileLock(UpperCAmelCase ): if os.path.exists(UpperCAmelCase ) and not args.overwrite_cache: lowerCamelCase__ : Tuple = time.time() lowerCamelCase__ : Union[str, Any] = torch.load(UpperCAmelCase ) logger.info( F"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) else: logger.info(F"""Creating features from dataset file at {args.data_dir}""" ) if mode == Split.dev: lowerCamelCase__ : str = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: lowerCamelCase__ : str = self.processor.get_test_examples(args.data_dir ) else: lowerCamelCase__ : List[str] = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: lowerCamelCase__ : Any = examples[:limit_length] lowerCamelCase__ : Tuple = glue_convert_examples_to_features( UpperCAmelCase , UpperCAmelCase , max_length=args.max_seq_length , label_list=UpperCAmelCase , output_mode=self.output_mode , ) lowerCamelCase__ : Union[str, Any] = time.time() torch.save(self.features , UpperCAmelCase ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self : Optional[int] ) -> Union[str, Any]: return len(self.features ) def __getitem__( self : Any , UpperCAmelCase : Optional[int] ) -> InputFeatures: return self.features[i] def A_ ( self : Tuple ) -> List[str]: return self.label_list
50
from itertools import count def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Optional[Any] = [1] * min_block_length for n in count(_UpperCAmelCase ): fill_count_functions.append(1 ) for block_length in range(_UpperCAmelCase , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) _UpperCAmelCase : List[str] = { """tanreinama/GPTSAN-2.8B-spout_is_uniform""": ( """https://huggingface.co/tanreinama/GPTSAN-2.8B-spout_is_uniform/resolve/main/config.json""" ), } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """gptsan-japanese""" UpperCAmelCase__ = [ """past_key_values""", ] UpperCAmelCase__ = { """hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : Optional[Any] , UpperCAmelCase : Optional[Any]=36000 , UpperCAmelCase : Union[str, Any]=1280 , UpperCAmelCase : int=1024 , UpperCAmelCase : Tuple=8192 , UpperCAmelCase : List[str]=4096 , UpperCAmelCase : Tuple=128 , UpperCAmelCase : Optional[Any]=10 , UpperCAmelCase : Dict=0 , UpperCAmelCase : Optional[int]=16 , UpperCAmelCase : str=16 , UpperCAmelCase : Optional[int]=128 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : Optional[int]=1e-5 , UpperCAmelCase : List[Any]=False , UpperCAmelCase : List[Any]=0.0 , UpperCAmelCase : int="float32" , UpperCAmelCase : int=False , UpperCAmelCase : Optional[Any]=False , UpperCAmelCase : Union[str, Any]=False , UpperCAmelCase : List[str]=0.0_0_2 , UpperCAmelCase : Optional[int]=False , UpperCAmelCase : Any=True , UpperCAmelCase : Optional[Any]=35998 , UpperCAmelCase : Union[str, Any]=35995 , UpperCAmelCase : Dict=35999 , **UpperCAmelCase : Union[str, Any] , ) -> List[str]: lowerCamelCase__ : int = vocab_size lowerCamelCase__ : List[str] = max_position_embeddings lowerCamelCase__ : str = d_model lowerCamelCase__ : Dict = d_ff lowerCamelCase__ : str = d_ext lowerCamelCase__ : List[Any] = d_spout lowerCamelCase__ : Optional[Any] = num_switch_layers lowerCamelCase__ : str = num_ext_layers lowerCamelCase__ : Dict = num_switch_layers + num_ext_layers lowerCamelCase__ : Tuple = num_heads lowerCamelCase__ : Optional[int] = num_experts lowerCamelCase__ : List[Any] = expert_capacity lowerCamelCase__ : Optional[int] = dropout_rate lowerCamelCase__ : Optional[Any] = layer_norm_epsilon lowerCamelCase__ : Optional[Any] = router_bias lowerCamelCase__ : str = router_jitter_noise lowerCamelCase__ : List[Any] = router_dtype lowerCamelCase__ : str = router_ignore_padding_tokens lowerCamelCase__ : Dict = output_hidden_states lowerCamelCase__ : Optional[Any] = output_attentions lowerCamelCase__ : Union[str, Any] = initializer_factor lowerCamelCase__ : Tuple = output_router_logits lowerCamelCase__ : str = use_cache super().__init__( separator_token_id=UpperCAmelCase , pad_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase , )
50
from __future__ import annotations from typing import Any def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: create_state_space_tree(_UpperCAmelCase , [] , 0 ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> None: if index == len(_UpperCAmelCase ): print(_UpperCAmelCase ) return create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _UpperCAmelCase : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["""A""", """B""", """C"""]) generate_all_subsequences(seq)
50
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase _UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) _UpperCAmelCase : Any = { """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json""", """allenai/longformer-large-4096""": """https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json""", """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json""" ), } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """longformer""" def __init__( self : Any , UpperCAmelCase : Union[List[int], int] = 512 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 1 , UpperCAmelCase : int = 0 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 30522 , UpperCAmelCase : int = 768 , UpperCAmelCase : int = 12 , UpperCAmelCase : int = 12 , UpperCAmelCase : int = 3072 , UpperCAmelCase : str = "gelu" , UpperCAmelCase : float = 0.1 , UpperCAmelCase : float = 0.1 , UpperCAmelCase : int = 512 , UpperCAmelCase : int = 2 , UpperCAmelCase : float = 0.0_2 , UpperCAmelCase : float = 1e-12 , UpperCAmelCase : bool = False , **UpperCAmelCase : int , ) -> Union[str, Any]: super().__init__(pad_token_id=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : str = attention_window lowerCamelCase__ : Optional[int] = sep_token_id lowerCamelCase__ : Optional[Any] = bos_token_id lowerCamelCase__ : int = eos_token_id lowerCamelCase__ : Any = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : str = num_hidden_layers lowerCamelCase__ : int = num_attention_heads lowerCamelCase__ : List[str] = hidden_act lowerCamelCase__ : Any = intermediate_size lowerCamelCase__ : Optional[int] = hidden_dropout_prob lowerCamelCase__ : List[str] = attention_probs_dropout_prob lowerCamelCase__ : Tuple = max_position_embeddings lowerCamelCase__ : str = type_vocab_size lowerCamelCase__ : List[Any] = initializer_range lowerCamelCase__ : Union[str, Any] = layer_norm_eps lowerCamelCase__ : Optional[Any] = onnx_export class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[Any] , UpperCAmelCase : "PretrainedConfig" , UpperCAmelCase : str = "default" , UpperCAmelCase : "List[PatchingSpec]" = None ) -> Any: super().__init__(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Any = True @property def A_ ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCamelCase__ : int = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCamelCase__ : Optional[int] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('global_attention_mask', dynamic_axis), ] ) @property def A_ ( self : Dict ) -> Mapping[str, Mapping[int, str]]: lowerCamelCase__ : Any = super().outputs if self.task == "default": lowerCamelCase__ : List[Any] = {0: 'batch'} return outputs @property def A_ ( self : Optional[int] ) -> float: return 1e-4 @property def A_ ( self : str ) -> int: # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def A_ ( self : List[str] , UpperCAmelCase : "PreTrainedTokenizerBase" , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: lowerCamelCase__ : List[str] = super().generate_dummy_inputs( preprocessor=UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly lowerCamelCase__ : Dict = torch.zeros_like(inputs['input_ids'] ) # make every second token global lowerCamelCase__ : Dict = 1 return inputs
50
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: _UpperCAmelCase : int = None _UpperCAmelCase : Dict = logging.get_logger(__name__) _UpperCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _UpperCAmelCase : List[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } _UpperCAmelCase : List[str] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off _UpperCAmelCase : Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] UpperCAmelCase__ = NllbTokenizer UpperCAmelCase__ = [] UpperCAmelCase__ = [] def __init__( self : Tuple , UpperCAmelCase : int=None , UpperCAmelCase : Any=None , UpperCAmelCase : str="<s>" , UpperCAmelCase : Optional[Any]="</s>" , UpperCAmelCase : str="</s>" , UpperCAmelCase : Tuple="<s>" , UpperCAmelCase : Optional[Any]="<unk>" , UpperCAmelCase : List[str]="<pad>" , UpperCAmelCase : Union[str, Any]="<mask>" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Any=False , **UpperCAmelCase : Optional[int] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : List[Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token lowerCamelCase__ : Union[str, Any] = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase , tokenizer_file=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , src_lang=UpperCAmelCase , tgt_lang=UpperCAmelCase , additional_special_tokens=UpperCAmelCase , legacy_behaviour=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : List[Any] = vocab_file lowerCamelCase__ : Dict = False if not self.vocab_file else True lowerCamelCase__ : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCamelCase__ : str = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase__ : int = src_lang if src_lang is not None else 'eng_Latn' lowerCamelCase__ : List[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A_ ( self : int ) -> str: return self._src_lang @src_lang.setter def A_ ( self : List[Any] , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A_ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Dict = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] , UpperCAmelCase : Optional[str] , **UpperCAmelCase : List[str] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCamelCase__ : Optional[int] = src_lang lowerCamelCase__ : Optional[int] = self(UpperCAmelCase , add_special_tokens=UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.convert_tokens_to_ids(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tgt_lang_id return inputs def A_ ( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : str = "eng_Latn" , UpperCAmelCase : Optional[List[str]] = None , UpperCAmelCase : str = "fra_Latn" , **UpperCAmelCase : Dict , ) -> BatchEncoding: lowerCamelCase__ : Any = src_lang lowerCamelCase__ : int = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Optional[int]: return self.set_src_lang_special_tokens(self.src_lang ) def A_ ( self : Any ) -> Union[str, Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A_ ( self : str , UpperCAmelCase : Optional[Any] ) -> None: lowerCamelCase__ : int = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : int = [] lowerCamelCase__ : str = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : int = [self.cur_lang_code] lowerCamelCase__ : Tuple = [self.eos_token_id] lowerCamelCase__ : Any = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : int , UpperCAmelCase : str ) -> None: lowerCamelCase__ : Union[str, Any] = self.convert_tokens_to_ids(UpperCAmelCase ) if self.legacy_behaviour: lowerCamelCase__ : Dict = [] lowerCamelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] else: lowerCamelCase__ : Any = [self.cur_lang_code] lowerCamelCase__ : Optional[Any] = [self.eos_token_id] lowerCamelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return lowerCamelCase__ : int = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase ): copyfile(self.vocab_file , UpperCAmelCase ) return (out_vocab_file,)
50
1
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> float: _validate_point(_UpperCAmelCase ) _validate_point(_UpperCAmelCase ) if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(a - b ) for a, b in zip(_UpperCAmelCase , _UpperCAmelCase ) ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> None: if point: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): for item in point: if not isinstance(_UpperCAmelCase , (int, float) ): lowerCamelCase__ : str = ( 'Expected a list of numbers as input, found ' F"""{type(_UpperCAmelCase ).__name__}""" ) raise TypeError(_UpperCAmelCase ) else: lowerCamelCase__ : Optional[Any] = F"""Expected a list of numbers as input, found {type(_UpperCAmelCase ).__name__}""" raise TypeError(_UpperCAmelCase ) else: raise ValueError('Missing an input' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> float: _validate_point(_UpperCAmelCase ) _validate_point(_UpperCAmelCase ) if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(x - y ) for x, y in zip(_UpperCAmelCase , _UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
50
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ , lowerCamelCase__ : List[str] = emb.weight.shape lowerCamelCase__ : Tuple = nn.Linear(_UpperCAmelCase , _UpperCAmelCase , bias=_UpperCAmelCase ) lowerCamelCase__ : Dict = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location='cpu' ) lowerCamelCase__ : List[str] = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCamelCase__ : Optional[int] = mam_aaa['model'] remove_ignore_keys_(_UpperCAmelCase ) lowerCamelCase__ : str = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCamelCase__ : Union[str, Any] = MaMaaaConfig( vocab_size=_UpperCAmelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCamelCase__ : Optional[Any] = state_dict['decoder.embed_tokens.weight'] lowerCamelCase__ : Union[str, Any] = MaMaaaForConditionalGeneration(_UpperCAmelCase ) model.model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) lowerCamelCase__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase : str = parser.parse_args() _UpperCAmelCase : Optional[Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
50
1
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[Any] , *UpperCAmelCase : Any , **UpperCAmelCase : int ) -> None: warnings.warn( 'The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DeiTImageProcessor instead.' , UpperCAmelCase , ) super().__init__(*UpperCAmelCase , **UpperCAmelCase )
50
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : str = set() lowerCamelCase__ : Any = [] def parse_line(_UpperCAmelCase ): for line in fp: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Any = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(_UpperCAmelCase ) > 0: lowerCamelCase__ : str = '\n'.join(_UpperCAmelCase ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(_UpperCAmelCase ) buffer.clear() continue else: lowerCamelCase__ : List[str] = line.strip() buffer.append(_UpperCAmelCase ) if from_gh: for filename in os.listdir(_UpperCAmelCase ): lowerCamelCase__ : Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) else: try: with zipfile.ZipFile(_UpperCAmelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_UpperCAmelCase ): # read the file if filename != "warnings.txt": continue with z.open(_UpperCAmelCase ) as fp: parse_line(_UpperCAmelCase ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = set() lowerCamelCase__ : Optional[int] = [os.path.join(_UpperCAmelCase , _UpperCAmelCase ) for p in os.listdir(_UpperCAmelCase ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_UpperCAmelCase , _UpperCAmelCase ) ) return selected_warnings if __name__ == "__main__": def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: return values.split(',' ) _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCAmelCase : Union[str, Any] = parser.parse_args() _UpperCAmelCase : Dict = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCAmelCase : Union[str, Any] = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCAmelCase : Dict = extract_warnings(args.output_dir, args.targets) _UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
50
1
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : List[Any] , UpperCAmelCase : NestedDataStructureLike[PathLike] , UpperCAmelCase : Optional[NamedSplit] = None , UpperCAmelCase : Optional[Features] = None , UpperCAmelCase : str = None , UpperCAmelCase : bool = False , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[int] = None , **UpperCAmelCase : Union[str, Any] , ) -> Tuple: super().__init__( UpperCAmelCase , split=UpperCAmelCase , features=UpperCAmelCase , cache_dir=UpperCAmelCase , keep_in_memory=UpperCAmelCase , streaming=UpperCAmelCase , num_proc=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : Union[str, Any] = path_or_paths if isinstance(UpperCAmelCase , UpperCAmelCase ) else {self.split: path_or_paths} lowerCamelCase__ : Dict = Text( cache_dir=UpperCAmelCase , data_files=UpperCAmelCase , features=UpperCAmelCase , **UpperCAmelCase , ) def A_ ( self : Optional[int] ) -> Optional[int]: # Build iterable dataset if self.streaming: lowerCamelCase__ : List[str] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Optional[Any] = None self.builder.download_and_prepare( download_config=UpperCAmelCase , download_mode=UpperCAmelCase , verification_mode=UpperCAmelCase , base_path=UpperCAmelCase , num_proc=self.num_proc , ) lowerCamelCase__ : List[Any] = self.builder.as_dataset( split=self.split , verification_mode=UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset
50
import flax.linen as nn import jax import jax.numpy as jnp class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : Any ) -> Any: lowerCamelCase__ : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , UpperCAmelCase : Dict ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = hidden_states.shape lowerCamelCase__ : Union[str, Any] = jax.image.resize( UpperCAmelCase , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> int: lowerCamelCase__ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : str , UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) lowerCamelCase__ : Optional[Any] = self.conv(UpperCAmelCase ) return hidden_states class lowerCAmelCase ( nn.Module ): UpperCAmelCase__ = 42 UpperCAmelCase__ = None UpperCAmelCase__ = 0.0 UpperCAmelCase__ = None UpperCAmelCase__ = jnp.floataa def A_ ( self : List[str] ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels lowerCamelCase__ : Tuple = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : int = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Union[str, Any] = nn.Dense(UpperCAmelCase , dtype=self.dtype ) lowerCamelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) lowerCamelCase__ : List[Any] = nn.Dropout(self.dropout_prob ) lowerCamelCase__ : Tuple = nn.Conv( UpperCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowerCamelCase__ : Optional[Any] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCamelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCamelCase__ : Dict = nn.Conv( UpperCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : int=True ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = hidden_states lowerCamelCase__ : List[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[Any] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Any = self.conva(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = self.time_emb_proj(nn.swish(UpperCAmelCase ) ) lowerCamelCase__ : List[str] = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase , 1 ) , 1 ) lowerCamelCase__ : List[str] = hidden_states + temb lowerCamelCase__ : Optional[Any] = self.norma(UpperCAmelCase ) lowerCamelCase__ : List[str] = nn.swish(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.dropout(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : str = self.conva(UpperCAmelCase ) if self.conv_shortcut is not None: lowerCamelCase__ : Dict = self.conv_shortcut(UpperCAmelCase ) return hidden_states + residual
50
1
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = 10 lowerCamelCase__ : int = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) lowerCamelCase__ : str = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(_UpperCAmelCase ) ), } , features=_UpperCAmelCase , ) return dataset @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ : Dict = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=_UpperCAmelCase ) return filename # FILE_CONTENT + files _UpperCAmelCase : Optional[int] = """\ Text data. Second line of data.""" @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt' lowerCamelCase__ : Optional[int] = FILE_CONTENT with open(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase ) return filename @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[Any]: import bza lowerCamelCase__ : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.bz2' lowerCamelCase__ : Any = bytes(_UpperCAmelCase , 'utf-8' ) with bza.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: import gzip lowerCamelCase__ : Any = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) lowerCamelCase__ : Optional[Any] = bytes(_UpperCAmelCase , 'utf-8' ) with gzip.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: if datasets.config.LZ4_AVAILABLE: import lza.frame lowerCamelCase__ : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.lz4' lowerCamelCase__ : int = bytes(_UpperCAmelCase , 'utf-8' ) with lza.frame.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: if datasets.config.PY7ZR_AVAILABLE: import pyazr lowerCamelCase__ : List[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.7z' with pyazr.SevenZipFile(_UpperCAmelCase , 'w' ) as archive: archive.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: import tarfile lowerCamelCase__ : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.tar' with tarfile.TarFile(_UpperCAmelCase , 'w' ) as f: f.add(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: import lzma lowerCamelCase__ : int = tmp_path_factory.mktemp('data' ) / 'file.txt.xz' lowerCamelCase__ : Optional[Any] = bytes(_UpperCAmelCase , 'utf-8' ) with lzma.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: import zipfile lowerCamelCase__ : Optional[int] = tmp_path_factory.mktemp('data' ) / 'file.txt.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd lowerCamelCase__ : Tuple = tmp_path_factory.mktemp('data' ) / 'file.txt.zst' lowerCamelCase__ : int = bytes(_UpperCAmelCase , 'utf-8' ) with zstd.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ : Optional[int] = tmp_path_factory.mktemp('data' ) / 'file.xml' lowerCamelCase__ : Optional[int] = textwrap.dedent( '\\n <?xml version="1.0" encoding="UTF-8" ?>\n <tmx version="1.4">\n <header segtype="sentence" srclang="ca" />\n <body>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang="en"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang="en"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang="en"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang="en"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang="en"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase ) return filename _UpperCAmelCase : Union[str, Any] = [ {"""col_1""": """0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """3""", """col_2""": 3, """col_3""": 3.0}, ] _UpperCAmelCase : List[Any] = [ {"""col_1""": """4""", """col_2""": 4, """col_3""": 4.0}, {"""col_1""": """5""", """col_2""": 5, """col_3""": 5.0}, ] _UpperCAmelCase : Optional[int] = { """col_1""": ["""0""", """1""", """2""", """3"""], """col_2""": [0, 1, 2, 3], """col_3""": [0.0, 1.0, 2.0, 3.0], } _UpperCAmelCase : Any = [ {"""col_3""": 0.0, """col_1""": """0""", """col_2""": 0}, {"""col_3""": 1.0, """col_1""": """1""", """col_2""": 1}, ] _UpperCAmelCase : int = [ {"""col_1""": """s0""", """col_2""": 0, """col_3""": 0.0}, {"""col_1""": """s1""", """col_2""": 1, """col_3""": 1.0}, {"""col_1""": """s2""", """col_2""": 2, """col_3""": 2.0}, {"""col_1""": """s3""", """col_2""": 3, """col_3""": 3.0}, ] @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( ) -> Optional[int]: return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : int = datasets.Dataset.from_dict(_UpperCAmelCase ) lowerCamelCase__ : str = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(_UpperCAmelCase ) ) as con: lowerCamelCase__ : List[str] = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(_UpperCAmelCase , 'w' , newline='' ) as f: lowerCamelCase__ : str = csv.DictWriter(_UpperCAmelCase , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(_UpperCAmelCase , 'w' , newline='' ) as f: lowerCamelCase__ : str = csv.DictWriter(_UpperCAmelCase , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> str: import bza lowerCamelCase__ : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.csv.bz2' with open(_UpperCAmelCase , 'rb' ) as f: lowerCamelCase__ : List[Any] = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(_UpperCAmelCase , 'wb' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: lowerCamelCase__ : int = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Dict: lowerCamelCase__ : List[str] = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(_UpperCAmelCase , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.csv.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) lowerCamelCase__ : List[str] = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(_UpperCAmelCase , 'wb' ) as f: lowerCamelCase__ : Union[str, Any] = pq.ParquetWriter(_UpperCAmelCase , schema=_UpperCAmelCase ) lowerCamelCase__ : Optional[int] = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(_UpperCAmelCase ) )] for k in DATA[0]} , schema=_UpperCAmelCase ) writer.write_table(_UpperCAmelCase ) writer.close() return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) lowerCamelCase__ : Tuple = {'data': DATA} with open(_UpperCAmelCase , 'w' ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) lowerCamelCase__ : int = {'data': DATA_DICT_OF_LISTS} with open(_UpperCAmelCase , 'w' ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Union[str, Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(_UpperCAmelCase , 'w' ) as f: for item in DATA: f.write(json.dumps(_UpperCAmelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[Any]: lowerCamelCase__ : str = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(_UpperCAmelCase , 'w' ) as f: for item in DATA: f.write(json.dumps(_UpperCAmelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : str = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(_UpperCAmelCase , 'w' ) as f: for item in DATA_312: f.write(json.dumps(_UpperCAmelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : str = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(_UpperCAmelCase , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(_UpperCAmelCase ) + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: import gzip lowerCamelCase__ : Union[str, Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(_UpperCAmelCase , 'rb' ) as orig_file: with gzip.open(_UpperCAmelCase , 'wb' ) as zipped_file: zipped_file.writelines(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: import gzip lowerCamelCase__ : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(_UpperCAmelCase , 'rb' ) as orig_file: with gzip.open(_UpperCAmelCase , 'wb' ) as zipped_file: zipped_file.writelines(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: lowerCamelCase__ : List[str] = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.join('nested' , os.path.basename(_UpperCAmelCase ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.jsonl.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.tar' with tarfile.TarFile(_UpperCAmelCase , 'w' ) as f: f.add(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) f.add(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: lowerCamelCase__ : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.tar' with tarfile.TarFile(_UpperCAmelCase , 'w' ) as f: f.add(_UpperCAmelCase , arcname=os.path.join('nested' , os.path.basename(_UpperCAmelCase ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : int = ['0', '1', '2', '3'] lowerCamelCase__ : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(_UpperCAmelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Dict = ['0', '1', '2', '3'] lowerCamelCase__ : Union[str, Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(_UpperCAmelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : List[Any] = ['0', '1', '2', '3'] lowerCamelCase__ : str = tmp_path_factory.mktemp('data' ) / 'dataset.abc' with open(_UpperCAmelCase , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.text.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.text.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) f.write(_UpperCAmelCase , arcname=os.path.join('main_dir' , os.path.basename(_UpperCAmelCase ) ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : List[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.ext.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename('unsupported.ext' ) ) f.write(_UpperCAmelCase , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Optional[int] = '\n'.join(['First', 'Second\u2029with Unicode new line', 'Third'] ) lowerCamelCase__ : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(_UpperCAmelCase ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( ) -> Tuple: return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( ) -> Dict: return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : int = tmp_path_factory.mktemp('data' ) / 'dataset.img.zip' with zipfile.ZipFile(_UpperCAmelCase , 'w' ) as f: f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ) ) f.write(_UpperCAmelCase , arcname=os.path.basename(_UpperCAmelCase ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Any: lowerCamelCase__ : List[Any] = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Optional[Any] = set() # edges = list of graph's edges lowerCamelCase__ : List[str] = get_edges(_UpperCAmelCase ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowerCamelCase__ , lowerCamelCase__ : str = edges.pop() chosen_vertices.add(_UpperCAmelCase ) chosen_vertices.add(_UpperCAmelCase ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(_UpperCAmelCase ) return chosen_vertices def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> set: lowerCamelCase__ : Union[str, Any] = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
50
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : List[str] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Dict = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: lowerCamelCase__ : List[Any] = [144, 192, 240] lowerCamelCase__ : Optional[Any] = [16, 32, 64, 96, 128, 160, 640] elif "mobilevit_xs" in mobilevit_name: lowerCamelCase__ : Dict = [96, 120, 144] lowerCamelCase__ : int = [16, 32, 48, 64, 80, 96, 384] elif "mobilevit_xxs" in mobilevit_name: lowerCamelCase__ : Tuple = [64, 80, 96] lowerCamelCase__ : List[str] = [16, 16, 24, 48, 64, 80, 320] lowerCamelCase__ : Union[str, Any] = 0.05 lowerCamelCase__ : Tuple = 2.0 if mobilevit_name.startswith('deeplabv3_' ): lowerCamelCase__ : Union[str, Any] = 512 lowerCamelCase__ : Optional[Any] = 16 lowerCamelCase__ : Dict = 21 lowerCamelCase__ : int = 'pascal-voc-id2label.json' else: lowerCamelCase__ : List[Any] = 1000 lowerCamelCase__ : int = 'imagenet-1k-id2label.json' lowerCamelCase__ : List[Any] = 'huggingface/label-files' lowerCamelCase__ : List[str] = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) , 'r' ) ) lowerCamelCase__ : Union[str, Any] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : List[str] = idalabel lowerCamelCase__ : Union[str, Any] = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=False ) -> List[str]: for i in range(1 , 6 ): if F"""layer_{i}.""" in name: lowerCamelCase__ : str = name.replace(F"""layer_{i}.""" , F"""encoder.layer.{i - 1}.""" ) if "conv_1." in name: lowerCamelCase__ : Any = name.replace('conv_1.' , 'conv_stem.' ) if ".block." in name: lowerCamelCase__ : Any = name.replace('.block.' , '.' ) if "exp_1x1" in name: lowerCamelCase__ : str = name.replace('exp_1x1' , 'expand_1x1' ) if "red_1x1" in name: lowerCamelCase__ : List[Any] = name.replace('red_1x1' , 'reduce_1x1' ) if ".local_rep.conv_3x3." in name: lowerCamelCase__ : Tuple = name.replace('.local_rep.conv_3x3.' , '.conv_kxk.' ) if ".local_rep.conv_1x1." in name: lowerCamelCase__ : str = name.replace('.local_rep.conv_1x1.' , '.conv_1x1.' ) if ".norm." in name: lowerCamelCase__ : str = name.replace('.norm.' , '.normalization.' ) if ".conv." in name: lowerCamelCase__ : List[Any] = name.replace('.conv.' , '.convolution.' ) if ".conv_proj." in name: lowerCamelCase__ : Dict = name.replace('.conv_proj.' , '.conv_projection.' ) for i in range(0 , 2 ): for j in range(0 , 4 ): if F""".{i}.{j}.""" in name: lowerCamelCase__ : Any = name.replace(F""".{i}.{j}.""" , F""".{i}.layer.{j}.""" ) for i in range(2 , 6 ): for j in range(0 , 4 ): if F""".{i}.{j}.""" in name: lowerCamelCase__ : Optional[int] = name.replace(F""".{i}.{j}.""" , F""".{i}.""" ) if "expand_1x1" in name: lowerCamelCase__ : int = name.replace('expand_1x1' , 'downsampling_layer.expand_1x1' ) if "conv_3x3" in name: lowerCamelCase__ : Union[str, Any] = name.replace('conv_3x3' , 'downsampling_layer.conv_3x3' ) if "reduce_1x1" in name: lowerCamelCase__ : Union[str, Any] = name.replace('reduce_1x1' , 'downsampling_layer.reduce_1x1' ) for i in range(2 , 5 ): if F""".global_rep.{i}.weight""" in name: lowerCamelCase__ : Optional[int] = name.replace(F""".global_rep.{i}.weight""" , '.layernorm.weight' ) if F""".global_rep.{i}.bias""" in name: lowerCamelCase__ : List[Any] = name.replace(F""".global_rep.{i}.bias""" , '.layernorm.bias' ) if ".global_rep." in name: lowerCamelCase__ : List[Any] = name.replace('.global_rep.' , '.transformer.' ) if ".pre_norm_mha.0." in name: lowerCamelCase__ : str = name.replace('.pre_norm_mha.0.' , '.layernorm_before.' ) if ".pre_norm_mha.1.out_proj." in name: lowerCamelCase__ : Union[str, Any] = name.replace('.pre_norm_mha.1.out_proj.' , '.attention.output.dense.' ) if ".pre_norm_ffn.0." in name: lowerCamelCase__ : str = name.replace('.pre_norm_ffn.0.' , '.layernorm_after.' ) if ".pre_norm_ffn.1." in name: lowerCamelCase__ : Optional[int] = name.replace('.pre_norm_ffn.1.' , '.intermediate.dense.' ) if ".pre_norm_ffn.4." in name: lowerCamelCase__ : Union[str, Any] = name.replace('.pre_norm_ffn.4.' , '.output.dense.' ) if ".transformer." in name: lowerCamelCase__ : str = name.replace('.transformer.' , '.transformer.layer.' ) if ".aspp_layer." in name: lowerCamelCase__ : Union[str, Any] = name.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in name: lowerCamelCase__ : Tuple = name.replace('.aspp_pool.' , '.' ) if "seg_head." in name: lowerCamelCase__ : Union[str, Any] = name.replace('seg_head.' , 'segmentation_head.' ) if "segmentation_head.classifier.classifier." in name: lowerCamelCase__ : Dict = name.replace('segmentation_head.classifier.classifier.' , 'segmentation_head.classifier.' ) if "classifier.fc." in name: lowerCamelCase__ : Optional[int] = name.replace('classifier.fc.' , 'classifier.' ) elif (not base_model) and ("segmentation_head." not in name): lowerCamelCase__ : Tuple = 'mobilevit.' + name return name def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ) -> Tuple: if base_model: lowerCamelCase__ : Any = '' else: lowerCamelCase__ : Any = 'mobilevit.' for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Any = orig_state_dict.pop(_UpperCAmelCase ) if key[:8] == "encoder.": lowerCamelCase__ : Union[str, Any] = key[8:] if "qkv" in key: lowerCamelCase__ : Optional[int] = key.split('.' ) lowerCamelCase__ : List[Any] = int(key_split[0][6:] ) - 1 lowerCamelCase__ : Tuple = int(key_split[3] ) lowerCamelCase__ : str = model.get_submodule(F"""{model_prefix}encoder.layer.{layer_num}""" ) lowerCamelCase__ : str = layer.transformer.layer[transformer_num].attention.attention.all_head_size lowerCamelCase__ : Union[str, Any] = ( F"""{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.""" ) if "weight" in key: lowerCamelCase__ : Any = val[:dim, :] lowerCamelCase__ : Dict = val[dim : dim * 2, :] lowerCamelCase__ : Any = val[-dim:, :] else: lowerCamelCase__ : Tuple = val[:dim] lowerCamelCase__ : List[Any] = val[dim : dim * 2] lowerCamelCase__ : List[str] = val[-dim:] else: lowerCamelCase__ : Union[str, Any] = val return orig_state_dict def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : List[str] = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ) -> str: lowerCamelCase__ : str = get_mobilevit_config(_UpperCAmelCase ) # load original state_dict lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location='cpu' ) # load 🤗 model if mobilevit_name.startswith('deeplabv3_' ): lowerCamelCase__ : Optional[Any] = MobileViTForSemanticSegmentation(_UpperCAmelCase ).eval() else: lowerCamelCase__ : Optional[int] = MobileViTForImageClassification(_UpperCAmelCase ).eval() lowerCamelCase__ : Optional[int] = convert_state_dict(_UpperCAmelCase , _UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) # Check outputs on an image, prepared by MobileViTImageProcessor lowerCamelCase__ : Any = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) lowerCamelCase__ : int = image_processor(images=prepare_img() , return_tensors='pt' ) lowerCamelCase__ : Union[str, Any] = model(**_UpperCAmelCase ) lowerCamelCase__ : int = outputs.logits if mobilevit_name.startswith('deeplabv3_' ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": lowerCamelCase__ : Any = torch.tensor( [ [[6.2_065, 6.1_292, 6.2_070], [6.1_079, 6.1_254, 6.1_747], [6.0_042, 6.1_071, 6.1_034]], [[-6.9_253, -6.8_653, -7.0_398], [-7.3_218, -7.3_983, -7.3_670], [-7.1_961, -7.2_482, -7.1_569]], [[-4.4_723, -4.4_348, -4.3_769], [-5.3_629, -5.4_632, -5.4_598], [-5.1_587, -5.3_402, -5.5_059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": lowerCamelCase__ : Tuple = torch.tensor( [ [[5.4_449, 5.5_733, 5.6_314], [5.1_815, 5.3_930, 5.5_963], [5.1_656, 5.4_333, 5.4_853]], [[-9.4_423, -9.7_766, -9.6_714], [-9.1_581, -9.5_720, -9.5_519], [-9.1_006, -9.6_458, -9.5_703]], [[-7.7_721, -7.3_716, -7.1_583], [-8.4_599, -8.0_624, -7.7_944], [-8.4_172, -7.8_366, -7.5_025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": lowerCamelCase__ : Optional[int] = torch.tensor( [ [[6.9_811, 6.9_743, 7.3_123], [7.1_777, 7.1_931, 7.3_938], [7.5_633, 7.8_050, 7.8_901]], [[-10.5_536, -10.2_332, -10.2_924], [-10.2_336, -9.8_624, -9.5_964], [-10.8_840, -10.8_158, -10.6_659]], [[-3.4_938, -3.0_631, -2.8_620], [-3.4_205, -2.8_135, -2.6_875], [-3.4_179, -2.7_945, -2.8_750]], ] ) else: raise ValueError(F"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3, :3, :3] , _UpperCAmelCase , atol=1e-4 ) else: assert logits.shape == (1, 1000) if mobilevit_name == "mobilevit_s": lowerCamelCase__ : Optional[int] = torch.tensor([-0.9_866, 0.2_392, -1.1_241] ) elif mobilevit_name == "mobilevit_xs": lowerCamelCase__ : Tuple = torch.tensor([-2.4_761, -0.9_399, -1.9_587] ) elif mobilevit_name == "mobilevit_xxs": lowerCamelCase__ : str = torch.tensor([-1.9_364, -1.2_327, -0.4_653] ) else: raise ValueError(F"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3] , _UpperCAmelCase , atol=1e-4 ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(F"""Saving model {mobilevit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: lowerCamelCase__ : Union[str, Any] = { 'mobilevit_s': 'mobilevit-small', 'mobilevit_xs': 'mobilevit-x-small', 'mobilevit_xxs': 'mobilevit-xx-small', 'deeplabv3_mobilevit_s': 'deeplabv3-mobilevit-small', 'deeplabv3_mobilevit_xs': 'deeplabv3-mobilevit-x-small', 'deeplabv3_mobilevit_xxs': 'deeplabv3-mobilevit-xx-small', } print('Pushing to the hub...' ) lowerCamelCase__ : Tuple = model_mapping[mobilevit_name] image_processor.push_to_hub(_UpperCAmelCase , organization='apple' ) model.push_to_hub(_UpperCAmelCase , organization='apple' ) if __name__ == "__main__": _UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--mobilevit_name""", default="""mobilevit_s""", type=str, help=( """Name of the MobileViT model you'd like to convert. Should be one of 'mobilevit_s', 'mobilevit_xs',""" """ 'mobilevit_xxs', 'deeplabv3_mobilevit_s', 'deeplabv3_mobilevit_xs', 'deeplabv3_mobilevit_xxs'.""" ), ) parser.add_argument( """--checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _UpperCAmelCase : Optional[int] = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
50
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
1
from itertools import count def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 50 ) -> int: lowerCamelCase__ : Optional[Any] = [1] * min_block_length for n in count(_UpperCAmelCase ): fill_count_functions.append(1 ) for block_length in range(_UpperCAmelCase , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
50
import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[str]: lowerCamelCase__ : str = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase__ : Optional[Any] = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase__ : List[str] = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase__ : Dict = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase__ : Tuple = key.replace(F"""patch_embed{idx}""" , F"""patch_embeddings.{int(_UpperCAmelCase )-1}""" ) if "norm" in key: lowerCamelCase__ : str = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase__ : Dict = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase__ : str = key.replace(F"""layer_norm{idx}""" , F"""layer_norm.{int(_UpperCAmelCase )-1}""" ) if "layer_norm1" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase__ : Optional[int] = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase__ : List[Any] = key[key.find('block' ) + len('block' )] lowerCamelCase__ : int = key.replace(F"""block{idx}""" , F"""block.{int(_UpperCAmelCase )-1}""" ) if "attn.q" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase__ : Union[str, Any] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: lowerCamelCase__ : Dict = key.replace('attn' , 'attention.self' ) if "fc1" in key: lowerCamelCase__ : Dict = key.replace('fc1' , 'dense1' ) if "fc2" in key: lowerCamelCase__ : Any = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: lowerCamelCase__ : Dict = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: lowerCamelCase__ : Tuple = key.replace('linear_fuse.conv' , 'linear_fuse' ) lowerCamelCase__ : List[str] = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase__ : Optional[Any] = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase__ : Dict = key.replace(F"""linear_c{idx}""" , F"""linear_c.{int(_UpperCAmelCase )-1}""" ) if "bot_conv" in key: lowerCamelCase__ : str = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: lowerCamelCase__ : Union[str, Any] = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: lowerCamelCase__ : List[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: lowerCamelCase__ : Optional[int] = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: lowerCamelCase__ : Union[str, Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: lowerCamelCase__ : List[Any] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase__ : str = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase__ : Dict = key.replace('module.last_layer_depth' , 'head.head' ) lowerCamelCase__ : str = value return new_state_dict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase__ : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase__ : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase__ : Optional[int] = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase__ : Optional[int] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase__ : Any = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase__ : Dict = kv_bias[config.hidden_sizes[i] :] def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase__ : Tuple = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Optional[int]: lowerCamelCase__ : str = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase__ : Union[str, Any] = GLPNImageProcessor() # prepare image lowerCamelCase__ : str = prepare_img() lowerCamelCase__ : Tuple = image_processor(images=_UpperCAmelCase , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase__ : Any = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase__ : str = rename_keys(_UpperCAmelCase ) # key and value matrices need special treatment read_in_k_v(_UpperCAmelCase , _UpperCAmelCase ) # create HuggingFace model and load state dict lowerCamelCase__ : Dict = GLPNForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # forward pass lowerCamelCase__ : List[str] = model(_UpperCAmelCase ) lowerCamelCase__ : Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase__ : List[Any] = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: lowerCamelCase__ : List[str] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase__ : Tuple = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _UpperCAmelCase , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_UpperCAmelCase , ) image_processor.push_to_hub( repo_path_or_name=Path(_UpperCAmelCase , _UpperCAmelCase ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_UpperCAmelCase , ) if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCAmelCase : int = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
50
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Dict , UpperCAmelCase : Union[str, Any]=99 , UpperCAmelCase : str=13 , UpperCAmelCase : List[str]=7 , UpperCAmelCase : str=9 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : str=True , UpperCAmelCase : Any=False , UpperCAmelCase : Union[str, Any]=32 , UpperCAmelCase : List[str]=5 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : Union[str, Any]=37 , UpperCAmelCase : int=8 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : Any=0.0_0_2 , UpperCAmelCase : Optional[Any]=1 , UpperCAmelCase : List[Any]=0 , UpperCAmelCase : Union[str, Any]=0 , UpperCAmelCase : Tuple=None , UpperCAmelCase : Optional[Any]=None , ) -> Union[str, Any]: lowerCamelCase__ : int = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Optional[int] = encoder_seq_length lowerCamelCase__ : int = decoder_seq_length # For common tests lowerCamelCase__ : List[str] = self.decoder_seq_length lowerCamelCase__ : Optional[int] = is_training lowerCamelCase__ : List[Any] = use_attention_mask lowerCamelCase__ : Optional[Any] = use_labels lowerCamelCase__ : Union[str, Any] = vocab_size lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : Optional[Any] = num_hidden_layers lowerCamelCase__ : Any = num_attention_heads lowerCamelCase__ : str = d_ff lowerCamelCase__ : Optional[Any] = relative_attention_num_buckets lowerCamelCase__ : Any = dropout_rate lowerCamelCase__ : Any = initializer_factor lowerCamelCase__ : Union[str, Any] = eos_token_id lowerCamelCase__ : List[str] = pad_token_id lowerCamelCase__ : List[str] = decoder_start_token_id lowerCamelCase__ : List[Any] = None lowerCamelCase__ : Optional[Any] = decoder_layers def A_ ( self : List[Any] ) -> int: return TaConfig.from_pretrained('google/umt5-base' ) def A_ ( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : Optional[int] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : Tuple=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : str=None , UpperCAmelCase : Optional[Any]=None , UpperCAmelCase : Optional[Any]=None , ) -> List[str]: if attention_mask is None: lowerCamelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCamelCase__ : Optional[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCamelCase__ : int = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase ) if decoder_head_mask is None: lowerCamelCase__ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) if cross_attn_head_mask is None: lowerCamelCase__ : Dict = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def A_ ( self : str ) -> List[str]: lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCamelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCamelCase__ : List[str] = input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Union[str, Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCamelCase__ : Dict = self.get_config() lowerCamelCase__ : Tuple = config.num_attention_heads lowerCamelCase__ : Any = self.prepare_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, input_dict def A_ ( self : Tuple ) -> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Dict = self.prepare_config_and_inputs() return config, inputs_dict def A_ ( self : Optional[int] ) -> List[str]: return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Union[str, Any] ) -> Dict: return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def A_ ( self : Any , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[str] , UpperCAmelCase : int , UpperCAmelCase : str , UpperCAmelCase : Any , UpperCAmelCase : Dict , ) -> str: lowerCamelCase__ : Dict = UMTaModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Optional[int] = model( input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , ) lowerCamelCase__ : Any = model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) lowerCamelCase__ : Dict = result.last_hidden_state lowerCamelCase__ : Any = result.past_key_values lowerCamelCase__ : List[Any] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def A_ ( self : Optional[int] , UpperCAmelCase : str , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : int , UpperCAmelCase : List[str] , UpperCAmelCase : Optional[Any] , ) -> Optional[int]: lowerCamelCase__ : List[Any] = UMTaModel(config=UpperCAmelCase ).get_decoder().to(UpperCAmelCase ).eval() # first forward pass lowerCamelCase__ : Tuple = model(UpperCAmelCase , use_cache=UpperCAmelCase ) lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : int = model(UpperCAmelCase , use_cache=UpperCAmelCase ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) ) self.parent.assertTrue(len(UpperCAmelCase ) == len(UpperCAmelCase ) + 1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCamelCase__ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase__ : List[str] = model(UpperCAmelCase )['last_hidden_state'] lowerCamelCase__ : str = model(UpperCAmelCase , past_key_values=UpperCAmelCase )['last_hidden_state'] # select random slice lowerCamelCase__ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase__ : Tuple = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase__ : List[str] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase , UpperCAmelCase , atol=1e-3 ) ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : List[str] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = UMTaModel(config=UpperCAmelCase ).to(UpperCAmelCase ).half().eval() lowerCamelCase__ : Optional[int] = model(**UpperCAmelCase )['last_hidden_state'] self.parent.assertFalse(torch.isnan(UpperCAmelCase ).any().item() ) @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) UpperCAmelCase__ = (UMTaForConditionalGeneration,) if is_torch_available() else () UpperCAmelCase__ = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = True UpperCAmelCase__ = True # The small UMT5 model needs higher percentages for CPU/MP tests UpperCAmelCase__ = [0.8, 0.9] def A_ ( self : Union[str, Any] ) -> List[Any]: lowerCamelCase__ : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def A_ ( self : Tuple ) -> int: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Tuple = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCAmelCase , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A_ ( self : Tuple ) -> Optional[Any]: lowerCamelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase ) def A_ ( self : List[Any] ) -> str: lowerCamelCase__ : int = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCamelCase__ : Any = config_and_inputs[0] lowerCamelCase__ : Any = UMTaForConditionalGeneration(UpperCAmelCase ).eval() model.to(UpperCAmelCase ) lowerCamelCase__ : Tuple = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ), } for attn_name, (name, mask) in zip(UpperCAmelCase , head_masking.items() ): lowerCamelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCamelCase__ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase ) lowerCamelCase__ : Tuple = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase , return_dict_in_generate=UpperCAmelCase , **UpperCAmelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCamelCase__ : Union[str, Any] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def A_ ( self : Optional[Any] ) -> Optional[Any]: pass @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=UpperCAmelCase ).to(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=UpperCAmelCase , legacy=UpperCAmelCase ) lowerCamelCase__ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] lowerCamelCase__ : Tuple = tokenizer(UpperCAmelCase , return_tensors='pt' , padding=UpperCAmelCase ).input_ids # fmt: off lowerCamelCase__ : Any = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Optional[int] = model.generate(input_ids.to(UpperCAmelCase ) ) lowerCamelCase__ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] lowerCamelCase__ : Union[str, Any] = tokenizer.batch_decode(UpperCAmelCase ) self.assertEqual(UpperCAmelCase , UpperCAmelCase )
50
1
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Callable: @wraps(_UpperCAmelCase ) def _inner_fn(*_UpperCAmelCase , **_UpperCAmelCase ): warnings.warn( (F"""'{fn.__name__}' is experimental and might be subject to breaking changes in the future.""") , _UpperCAmelCase , ) return fn(*_UpperCAmelCase , **_UpperCAmelCase ) return _inner_fn
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = 42 class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase ): @register_to_config def __init__( self : List[str] , UpperCAmelCase : int = 65536 , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 0 , UpperCAmelCase : str = "fourier" , UpperCAmelCase : bool = True , UpperCAmelCase : bool = False , UpperCAmelCase : float = 0.0 , UpperCAmelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase : str = None , UpperCAmelCase : Tuple[int] = (32, 32, 64) , UpperCAmelCase : str = None , UpperCAmelCase : int = 8 , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = False , ) -> List[Any]: super().__init__() lowerCamelCase__ : Optional[int] = sample_size # time if time_embedding_type == "fourier": lowerCamelCase__ : Optional[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase , log=UpperCAmelCase , flip_sin_to_cos=UpperCAmelCase ) lowerCamelCase__ : Any = 2 * block_out_channels[0] elif time_embedding_type == "positional": lowerCamelCase__ : List[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase , downscale_freq_shift=UpperCAmelCase ) lowerCamelCase__ : Dict = block_out_channels[0] if use_timestep_embedding: lowerCamelCase__ : str = block_out_channels[0] * 4 lowerCamelCase__ : List[Any] = TimestepEmbedding( in_channels=UpperCAmelCase , time_embed_dim=UpperCAmelCase , act_fn=UpperCAmelCase , out_dim=block_out_channels[0] , ) lowerCamelCase__ : Any = nn.ModuleList([] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : List[str] = nn.ModuleList([] ) lowerCamelCase__ : Optional[int] = None # down lowerCamelCase__ : Optional[int] = in_channels for i, down_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = output_channel lowerCamelCase__ : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels lowerCamelCase__ : Union[str, Any] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Optional[int] = get_down_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase ) # mid lowerCamelCase__ : Optional[int] = get_mid_block( UpperCAmelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase , add_downsample=UpperCAmelCase , ) # up lowerCamelCase__ : Optional[int] = list(reversed(UpperCAmelCase ) ) lowerCamelCase__ : Optional[int] = reversed_block_out_channels[0] if out_block_type is None: lowerCamelCase__ : List[str] = out_channels else: lowerCamelCase__ : Any = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase ): lowerCamelCase__ : Tuple = output_channel lowerCamelCase__ : Union[str, Any] = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase ) - 1 else final_upsample_channels ) lowerCamelCase__ : List[str] = i == len(UpperCAmelCase ) - 1 lowerCamelCase__ : Dict = get_up_block( UpperCAmelCase , num_layers=UpperCAmelCase , in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase ) lowerCamelCase__ : int = output_channel # out lowerCamelCase__ : int = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) lowerCamelCase__ : List[Any] = get_out_block( out_block_type=UpperCAmelCase , num_groups_out=UpperCAmelCase , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase , act_fn=UpperCAmelCase , fc_dim=block_out_channels[-1] // 4 , ) def A_ ( self : List[Any] , UpperCAmelCase : torch.FloatTensor , UpperCAmelCase : Union[torch.Tensor, float, int] , UpperCAmelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: lowerCamelCase__ : Optional[Any] = timestep if not torch.is_tensor(UpperCAmelCase ): lowerCamelCase__ : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase ) and len(timesteps.shape ) == 0: lowerCamelCase__ : List[str] = timesteps[None].to(sample.device ) lowerCamelCase__ : Optional[int] = self.time_proj(UpperCAmelCase ) if self.config.use_timestep_embedding: lowerCamelCase__ : str = self.time_mlp(UpperCAmelCase ) else: lowerCamelCase__ : List[str] = timestep_embed[..., None] lowerCamelCase__ : str = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) lowerCamelCase__ : str = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down lowerCamelCase__ : str = () for downsample_block in self.down_blocks: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = downsample_block(hidden_states=UpperCAmelCase , temb=UpperCAmelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: lowerCamelCase__ : Optional[Any] = self.mid_block(UpperCAmelCase , UpperCAmelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): lowerCamelCase__ : Dict = down_block_res_samples[-1:] lowerCamelCase__ : Optional[Any] = down_block_res_samples[:-1] lowerCamelCase__ : Any = upsample_block(UpperCAmelCase , res_hidden_states_tuple=UpperCAmelCase , temb=UpperCAmelCase ) # 5. post-process if self.out_block: lowerCamelCase__ : Any = self.out_block(UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase )
50
1
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCAmelCase : def __init__( self : Tuple , UpperCAmelCase : Any , UpperCAmelCase : List[Any]=13 , UpperCAmelCase : Dict=30 , UpperCAmelCase : Dict=2 , UpperCAmelCase : int=3 , UpperCAmelCase : Dict=True , UpperCAmelCase : str=True , UpperCAmelCase : Dict=32 , UpperCAmelCase : List[Any]=5 , UpperCAmelCase : str=4 , UpperCAmelCase : List[Any]=37 , UpperCAmelCase : Dict="gelu" , UpperCAmelCase : Optional[Any]=0.1 , UpperCAmelCase : Tuple=0.1 , UpperCAmelCase : List[Any]=10 , UpperCAmelCase : List[Any]=0.0_2 , UpperCAmelCase : Dict=3 , UpperCAmelCase : int=0.6 , UpperCAmelCase : Union[str, Any]=None , ) -> Any: lowerCamelCase__ : Tuple = parent lowerCamelCase__ : Tuple = batch_size lowerCamelCase__ : Any = image_size lowerCamelCase__ : int = patch_size lowerCamelCase__ : Dict = num_channels lowerCamelCase__ : Union[str, Any] = is_training lowerCamelCase__ : Union[str, Any] = use_labels lowerCamelCase__ : Union[str, Any] = hidden_size lowerCamelCase__ : str = num_hidden_layers lowerCamelCase__ : str = num_attention_heads lowerCamelCase__ : Tuple = intermediate_size lowerCamelCase__ : str = hidden_act lowerCamelCase__ : Optional[Any] = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : Any = type_sequence_label_size lowerCamelCase__ : Any = initializer_range lowerCamelCase__ : str = mask_ratio lowerCamelCase__ : Tuple = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase__ : Optional[int] = (image_size // patch_size) ** 2 lowerCamelCase__ : Dict = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def A_ ( self : List[Any] ) -> Tuple: lowerCamelCase__ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[str] = None if self.use_labels: lowerCamelCase__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ : List[Any] = self.get_config() return config, pixel_values, labels def A_ ( self : str ) -> Any: return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def A_ ( self : Optional[Any] , UpperCAmelCase : Tuple , UpperCAmelCase : Optional[Any] , UpperCAmelCase : int ) -> List[Any]: lowerCamelCase__ : Optional[Any] = ViTMAEModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : Any = model(UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self : List[str] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] ) -> Any: lowerCamelCase__ : List[Any] = ViTMAEForPreTraining(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCAmelCase ) lowerCamelCase__ : Any = (self.image_size // self.patch_size) ** 2 lowerCamelCase__ : str = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase__ : str = 1 lowerCamelCase__ : Union[str, Any] = ViTMAEForPreTraining(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() lowerCamelCase__ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ : str = model(UpperCAmelCase ) lowerCamelCase__ : str = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def A_ ( self : Dict ) -> Union[str, Any]: lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple = config_and_inputs lowerCamelCase__ : List[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( __UpperCamelCase, __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () UpperCAmelCase__ = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def A_ ( self : List[str] ) -> List[str]: lowerCamelCase__ : List[Any] = ViTMAEModelTester(self ) lowerCamelCase__ : Optional[Any] = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def A_ ( self : str ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def A_ ( self : Tuple ) -> Union[str, Any]: pass def A_ ( self : List[str] ) -> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[Any] = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCamelCase__ : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def A_ ( self : Optional[int] ) -> Optional[int]: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCAmelCase ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : str = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def A_ ( self : List[Any] ) -> Optional[int]: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def A_ ( self : Any ) -> int: lowerCamelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCAmelCase ) def A_ ( self : str , UpperCAmelCase : List[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : int ) -> str: # make masks reproducible np.random.seed(2 ) lowerCamelCase__ : List[str] = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) lowerCamelCase__ : Any = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase__ : List[str] = torch.from_numpy(UpperCAmelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase__ : List[Any] = pt_noise super().check_pt_tf_models(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def A_ ( self : List[Any] ) -> Optional[int]: lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowerCamelCase__ : Optional[Any] = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) lowerCamelCase__ : Union[str, Any] = outputs[0].cpu().numpy() lowerCamelCase__ : int = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCAmelCase ) lowerCamelCase__ : List[Any] = model_class.from_pretrained(UpperCAmelCase ) model.to(UpperCAmelCase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowerCamelCase__ : int = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) # Make sure we don't have nans lowerCamelCase__ : str = after_outputs[0].cpu().numpy() lowerCamelCase__ : List[Any] = 0 lowerCamelCase__ : List[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(UpperCAmelCase , 1e-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def A_ ( self : int ) -> Union[str, Any]: pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def A_ ( self : List[str] ) -> str: pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def A_ ( self : List[str] ) -> Any: pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def A_ ( self : Tuple ) -> Tuple: pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def A_ ( self : List[Any] ) -> Dict: pass @slow def A_ ( self : int ) -> Union[str, Any]: for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Optional[Any] = ViTMAEModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( ) -> List[str]: lowerCamelCase__ : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowerCAmelCase ( unittest.TestCase ): @cached_property def A_ ( self : Dict ) -> List[str]: return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def A_ ( self : str ) -> Tuple: # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase__ : str = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(UpperCAmelCase ) lowerCamelCase__ : Optional[int] = self.default_image_processor lowerCamelCase__ : Any = prepare_img() lowerCamelCase__ : Dict = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase__ : Union[str, Any] = ViTMAEConfig() lowerCamelCase__ : Dict = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase__ : List[Any] = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): lowerCamelCase__ : Any = model(**UpperCAmelCase , noise=torch.from_numpy(UpperCAmelCase ).to(device=UpperCAmelCase ) ) # verify the logits lowerCamelCase__ : Any = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) lowerCamelCase__ : str = torch.tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(UpperCAmelCase ) , atol=1e-4 ) )
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Dict = {"""configuration_sew""": ["""SEW_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SEWConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : str = [ """SEW_PRETRAINED_MODEL_ARCHIVE_LIST""", """SEWForCTC""", """SEWForSequenceClassification""", """SEWModel""", """SEWPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_sew import SEW_PRETRAINED_CONFIG_ARCHIVE_MAP, SEWConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_sew import ( SEW_PRETRAINED_MODEL_ARCHIVE_LIST, SEWForCTC, SEWForSequenceClassification, SEWModel, SEWPreTrainedModel, ) else: import sys _UpperCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
50
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : Optional[int] = [] embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", F"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", F"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", F"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( F"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", F"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Tuple = [] attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", F"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( F"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", F"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", F"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", F"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", F"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (F"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", F"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Union[str, Any] = [] token.append((F"""cvt.encoder.stages.{idx}.cls_token""", 'stage2.cls_token') ) return token def SCREAMING_SNAKE_CASE ( ) -> str: lowerCamelCase__ : str = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Tuple = 'imagenet-1k-id2label.json' lowerCamelCase__ : Union[str, Any] = 1000 lowerCamelCase__ : Optional[Any] = 'huggingface/label-files' lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Dict = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowerCamelCase__ : int = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Tuple = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : List[str] = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowerCamelCase__ : List[Any] = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowerCamelCase__ : Dict = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowerCamelCase__ : Optional[Any] = [2, 2, 20] lowerCamelCase__ : Optional[int] = [3, 12, 16] lowerCamelCase__ : str = [192, 768, 1024] lowerCamelCase__ : Any = CvtForImageClassification(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : List[str] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowerCamelCase__ : Optional[int] = OrderedDict() lowerCamelCase__ : Tuple = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowerCamelCase__ : Optional[Any] = list_of_state_dict + cls_token(_UpperCAmelCase ) lowerCamelCase__ : str = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowerCamelCase__ : str = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : int = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : str = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( """--cvt_model""", default="""cvt-w24""", type=str, help="""Name of the cvt model you'd like to convert.""", ) parser.add_argument( """--image_size""", default=3_84, type=int, help="""Input Image Size""", ) parser.add_argument( """--cvt_file_name""", default=R"""cvtmodels\CvT-w24-384x384-IN-22k.pth""", type=str, help="""Input Image Size""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCAmelCase : List[str] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
50
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[Any] , UpperCAmelCase : int ) -> int: lowerCamelCase__ : Tuple = data def __iter__( self : Dict ) -> List[str]: for element in self.data: yield element def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=True ) -> int: lowerCamelCase__ : Optional[int] = Accelerator(even_batches=_UpperCAmelCase ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = False ) -> List[str]: if iterable: lowerCamelCase__ : Optional[Any] = DummyIterableDataset(torch.as_tensor(range(_UpperCAmelCase ) ) ) else: lowerCamelCase__ : Tuple = TensorDataset(torch.as_tensor(range(_UpperCAmelCase ) ) ) lowerCamelCase__ : str = DataLoader(_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : str = accelerator.prepare(_UpperCAmelCase ) return dl def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> str: lowerCamelCase__ : Any = create_dataloader(accelerator=_UpperCAmelCase , dataset_size=_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : Any = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : List[str] = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : Tuple = create_accelerator(even_batches=_UpperCAmelCase ) verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: lowerCamelCase__ : Tuple = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Any = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : List[Any] = accelerator.prepare(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) lowerCamelCase__ : List[str] = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = ddp_model(batch[0].float() ) lowerCamelCase__ : int = output.sum() loss.backward() batch_idxs.append(_UpperCAmelCase ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: with warnings.catch_warnings(record=_UpperCAmelCase ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , _UpperCAmelCase ) assert "only supported for multi-GPU" in str(w[-1].message ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Tuple = False lowerCamelCase__ : List[str] = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : List[Any] = accelerator.prepare(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) lowerCamelCase__ : List[Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = train_dl.batch_sampler.even_batches lowerCamelCase__ : Union[str, Any] = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Optional[int] = False lowerCamelCase__ : Any = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Tuple = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : Tuple = accelerator.prepare(_UpperCAmelCase ) create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 , iterable=_UpperCAmelCase ) lowerCamelCase__ : int = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings('ignore' ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): lowerCamelCase__ : str = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : int = create_accelerator() lowerCamelCase__ : str = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : Any = accelerator.prepare(_UpperCAmelCase ) create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 , iterable=_UpperCAmelCase ) with warnings.catch_warnings(record=_UpperCAmelCase ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): pass assert issubclass(w[-1].category , _UpperCAmelCase ) assert "only supported for map-style datasets" in str(w[-1].message ) def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : str = create_accelerator() accelerator.print('Test that even_batches variable ensures uniform batches across processes' ) test_default_ensures_even_batch_sizes() accelerator.print('Run tests with even_batches disabled' ) test_can_disable_even_batches() accelerator.print('Test joining uneven inputs' ) test_can_join_uneven_inputs() accelerator.print('Test overriding even_batches when joining uneven inputs' ) test_join_can_override_even_batches() accelerator.print('Test overriding even_batches for mixed dataloader types' ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print('Test overriding even_batches raises a warning for iterable dataloaders' ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print('Test join with non DDP distributed raises warning' ) lowerCamelCase__ : Dict = accelerator.state.distributed_type lowerCamelCase__ : List[Any] = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(_UpperCAmelCase ) lowerCamelCase__ : Dict = original_state if __name__ == "__main__": main()
50
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=None ) -> Tuple: if subparsers is not None: lowerCamelCase__ : Any = subparsers.add_parser('test' ) else: lowerCamelCase__ : int = argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' , default=_UpperCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Union[str, Any]: lowerCamelCase__ : Tuple = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: lowerCamelCase__ : List[str] = script_name else: lowerCamelCase__ : List[Any] = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase__ : str = ['accelerate-launch'] + test_args.split() lowerCamelCase__ : Dict = execute_subprocess_async(_UpperCAmelCase , env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def SCREAMING_SNAKE_CASE ( ) -> Any: lowerCamelCase__ : Any = test_command_parser() lowerCamelCase__ : List[Any] = parser.parse_args() test_command(_UpperCAmelCase ) if __name__ == "__main__": main()
50
1
import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Tuple = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } _UpperCAmelCase : Any = { """vocab_file""": {"""facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"""}, """merges_file""": {"""facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"""}, """tokenizer_config_file""": { """facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json""" }, } _UpperCAmelCase : str = {"""facebook/blenderbot-3B""": 1_28} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def SCREAMING_SNAKE_CASE ( ) -> Optional[int]: lowerCamelCase__ : Optional[int] = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowerCamelCase__ : Optional[Any] = bs[:] lowerCamelCase__ : List[Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(_UpperCAmelCase ) cs.append(2**8 + n ) n += 1 lowerCamelCase__ : Any = [chr(_UpperCAmelCase ) for n in cs] return dict(zip(_UpperCAmelCase , _UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = set() lowerCamelCase__ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCamelCase__ : Dict = char return pairs class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any] , UpperCAmelCase : List[str] , UpperCAmelCase : Dict , UpperCAmelCase : Optional[Any]="replace" , UpperCAmelCase : str="<s>" , UpperCAmelCase : Tuple="</s>" , UpperCAmelCase : Dict="</s>" , UpperCAmelCase : Union[str, Any]="<s>" , UpperCAmelCase : int="<unk>" , UpperCAmelCase : Optional[Any]="<pad>" , UpperCAmelCase : int="<mask>" , UpperCAmelCase : str=False , **UpperCAmelCase : Any , ) -> Optional[Any]: lowerCamelCase__ : List[str] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else bos_token lowerCamelCase__ : Tuple = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else eos_token lowerCamelCase__ : Dict = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else sep_token lowerCamelCase__ : Union[str, Any] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else cls_token lowerCamelCase__ : Any = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else unk_token lowerCamelCase__ : List[str] = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase__ : str = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else mask_token super().__init__( errors=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , add_prefix_space=UpperCAmelCase , **UpperCAmelCase , ) with open(UpperCAmelCase , encoding='utf-8' ) as vocab_handle: lowerCamelCase__ : int = json.load(UpperCAmelCase ) lowerCamelCase__ : List[Any] = {v: k for k, v in self.encoder.items()} lowerCamelCase__ : Tuple = errors # how to handle errors in decoding lowerCamelCase__ : List[str] = bytes_to_unicode() lowerCamelCase__ : Union[str, Any] = {v: k for k, v in self.byte_encoder.items()} with open(UpperCAmelCase , encoding='utf-8' ) as merges_handle: lowerCamelCase__ : Any = merges_handle.read().split('\n' )[1:-1] lowerCamelCase__ : Union[str, Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCamelCase__ : Tuple = dict(zip(UpperCAmelCase , range(len(UpperCAmelCase ) ) ) ) lowerCamelCase__ : Dict = {} lowerCamelCase__ : Optional[Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCamelCase__ : Optional[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def A_ ( self : Any ) -> Optional[Any]: return len(self.encoder ) def A_ ( self : Optional[int] ) -> Optional[Any]: return dict(self.encoder , **self.added_tokens_encoder ) def A_ ( self : Optional[int] , UpperCAmelCase : Optional[int] ) -> Optional[Any]: if token in self.cache: return self.cache[token] lowerCamelCase__ : Union[str, Any] = tuple(UpperCAmelCase ) lowerCamelCase__ : int = get_pairs(UpperCAmelCase ) if not pairs: return token while True: lowerCamelCase__ : List[str] = min(UpperCAmelCase , key=lambda UpperCAmelCase : self.bpe_ranks.get(UpperCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowerCamelCase__ , lowerCamelCase__ : int = bigram lowerCamelCase__ : Optional[Any] = [] lowerCamelCase__ : Union[str, Any] = 0 while i < len(UpperCAmelCase ): try: lowerCamelCase__ : int = word.index(UpperCAmelCase , UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCamelCase__ : Optional[int] = j if word[i] == first and i < len(UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCamelCase__ : str = tuple(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = new_word if len(UpperCAmelCase ) == 1: break else: lowerCamelCase__ : Optional[Any] = get_pairs(UpperCAmelCase ) lowerCamelCase__ : int = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Tuple = word return word def A_ ( self : int , UpperCAmelCase : List[Any] ) -> Union[str, Any]: lowerCamelCase__ : int = [] for token in re.findall(self.pat , UpperCAmelCase ): lowerCamelCase__ : List[Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(UpperCAmelCase ).split(' ' ) ) return bpe_tokens def A_ ( self : Any , UpperCAmelCase : List[Any] ) -> List[Any]: return self.encoder.get(UpperCAmelCase , self.encoder.get(self.unk_token ) ) def A_ ( self : Union[str, Any] , UpperCAmelCase : Any ) -> str: return self.decoder.get(UpperCAmelCase ) def A_ ( self : Optional[Any] , UpperCAmelCase : int ) -> int: lowerCamelCase__ : str = ''.join(UpperCAmelCase ) lowerCamelCase__ : int = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def A_ ( self : str , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCamelCase__ : List[str] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowerCamelCase__ : Union[str, Any] = os.path.join( UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=UpperCAmelCase , ensure_ascii=UpperCAmelCase ) + '\n' ) lowerCamelCase__ : Any = 0 with open(UpperCAmelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowerCamelCase__ : Union[str, Any] = token_index writer.write(' '.join(UpperCAmelCase ) + '\n' ) index += 1 return vocab_file, merge_file def A_ ( self : Union[str, Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None , UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase , token_ids_a=UpperCAmelCase , already_has_special_tokens=UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase )) + [1] return [1] + ([0] * len(UpperCAmelCase )) + [1, 1] + ([0] * len(UpperCAmelCase )) + [1] def A_ ( self : str , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : Any = [self.sep_token_id] lowerCamelCase__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A_ ( self : int , UpperCAmelCase : str , UpperCAmelCase : Optional[Any]=False , **UpperCAmelCase : List[Any] ) -> List[Any]: lowerCamelCase__ : Any = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(UpperCAmelCase ) > 0 and not text[0].isspace()): lowerCamelCase__ : str = ' ' + text return (text, kwargs) def A_ ( self : List[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> str: return token_ids_a + [self.eos_token_id] def A_ ( self : Any , UpperCAmelCase : "Conversation" ) -> List[int]: lowerCamelCase__ : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(' ' + text ) else: # Generated responses should contain them already. inputs.append(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = ' '.join(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = self.encode(UpperCAmelCase ) if len(UpperCAmelCase ) > self.model_max_length: lowerCamelCase__ : List[str] = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
1